Get AI summaries of any video or article — Sign up free
Exploratory Factor Analysis in SPSS for Scale Development thumbnail

Exploratory Factor Analysis in SPSS for Scale Development

Research With Fawad·
5 min read

Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Build the initial item pool from both prior scales and qualitative interviews to address gaps in existing measures (e.g., business-focused instruments vs. university-focused concepts).

Briefing

Exploratory factor analysis (EFA) in SPSS is used to test whether a newly built multi-dimensional scale actually behaves the way theory predicts—specifically, whether items cluster into the intended factors. In this workflow, the scale targets “university social responsibility,” operationalized through three dimensions: ethical responsibilities, research and development responsibilities, and philanthropic responsibilities. The practical goal is straightforward: run EFA, check whether a three-factor structure emerges, and then prune items that fail to load cleanly.

The scale development process leading into EFA follows a two-source logic. First, existing literature is scanned to collect items used in prior measures of social responsibility, with an explicit effort to address a gap: many existing instruments focus on business organizations rather than universities. Second, focus group discussions and interviews add new candidate statements. Keywords from these qualitative sessions—such as the idea that universities should have a code of conduct—are converted into survey items. At the end of this stage, the construct sits at the top of a hierarchy: construct → dimensions → items, with theory predicting which items belong to which dimension.

EFA in SPSS begins by selecting the items under “Analyze → Dimension Reduction → Factor.” Several diagnostics are used to justify factor analysis. The Bartlett’s test of sphericity is checked for significance, indicating that correlations among variables are sufficiently patterned for factor extraction. The Kaiser–Meyer–Olkin (KMO) measure of sampling adequacy is also reported; here it reaches 0.931, signaling strong suitability. Extraction uses principal component analysis, and the analysis retains three factors based on the eigenvalue rule (factors with I value greater than one) and the cumulative explained variance, which lands at roughly 60% for the three-factor solution.

Rotation is then applied (Varimax rotation) to clarify which items load onto which factor. Initial results show that most ethical responsibility items load well together, and philanthropic responsibility items also form a coherent cluster. However, problems appear in the research and development dimension: one item (RD2) loads weakly or onto the wrong factor, and two items (RD1 and PR1) show no meaningful loading (below the 0.050 threshold). The workflow treats this as a signal to remove problematic items rather than forcing a messy structure. After removing RD2, then RD1, and then PR1—one by one, with re-runs to see how the structure stabilizes—the factor pattern becomes clean and matches the expected three-factor model.

Reporting the results is treated as part of the method, not an afterthought. The write-up should include the extraction method (principal component analysis), rotation (Varimax), the minimum factor loading criterion (0.050), and the Bartlett test and KMO values. Commonalities and model fit are also documented, including the proportion of non-redundant residuals (kept below 50% as an indicator of acceptable fit). Importantly, the process emphasizes that item deletion should be justified: items are removed because they fail to load on the intended factor or fail to load meaningfully at all. The final output is summarized in a factor-loading table with Factor 1, Factor 2, and Factor 3 labeled according to the three dimensions—ethical responsibilities, research and development responsibilities, and philanthropic responsibilities—producing a structure that aligns with the original theoretical expectations.

Cornell Notes

The scale for “university social responsibility” is built from literature items plus new statements generated through interviews and focus groups, then tested with exploratory factor analysis (EFA) in SPSS. Theory predicts three dimensions—ethical responsibilities, research and development responsibilities, and philanthropic responsibilities—so EFA should yield a three-factor solution with items clustering accordingly. Diagnostics support factorability: KMO is 0.931 and Bartlett’s test is significant. The initial three-factor extraction explains about 60% of variance, but several items fail the loading rules (below 0.050 or loading onto the wrong factor). Removing RD2, RD1, and PR1 produces a clearer, theory-consistent factor structure, which is then reported with KMO, Bartlett’s test, commonalities, and factor loadings.

How does the scale development process feed into EFA expectations?

Items originate from two sources: existing research scales (to address a gap where many measures target businesses rather than universities) and qualitative work (focus groups/interviews). Keywords from interviews—such as the need for a code of conduct—are converted into survey statements. Those items are then mapped into a theory-driven hierarchy: construct → three dimensions → items, with the expectation that EFA will show items loading together within each dimension.

What SPSS checks justify running EFA on the item set?

Before interpreting factors, the workflow checks Bartlett’s test of sphericity (significant results indicate correlations are suitable for factor analysis) and the Kaiser–Meyer–Olkin (KMO) measure (here KMO = 0.931, indicating strong sampling adequacy). The analysis also inspects commonalities and model fit indicators such as the percentage of non-redundant residuals (kept under 50% for acceptable fit).

Why was a three-factor solution retained, and what evidence supported it?

Extraction uses principal component analysis, and the retention rule keeps factors with eigenvalues (I value) greater than one. The output also shows a cumulative explained variance close to 60% for three factors, supporting the decision not to extract additional factors beyond the intended three.

What loading problems triggered item deletion?

After Varimax rotation, most ethical and philanthropic items load well above the minimum threshold (0.050). In contrast, research and development items show issues: RD2 loads weakly or on the wrong factor, RD1 shows no loading (below 0.050), and PR1 also shows no loading. Because these items do not represent their intended dimension, they are removed.

How did the factor structure change after removing items?

The process removes items one at a time and re-runs EFA to see whether the structure stabilizes. After deleting RD2, then RD1, then PR1, the rotated component matrix becomes clear: three factors emerge cleanly and align with the expected dimensions—ethical responsibilities, research and development responsibilities, and philanthropic responsibilities.

What should be included when reporting EFA results in a write-up?

A complete report includes: the extraction method (principal component analysis), rotation (Varimax), the minimum factor loading criterion (0.050), Bartlett’s test and KMO, commonalities (noting any items slightly below the commonality threshold but still loading well), and model-fit details such as non-redundant residuals. It also documents which items were removed and why (wrong-factor loading or no meaningful loading), then presents the final factor-loading table with factors labeled by dimension.

Review Questions

  1. If an item loads below 0.050 on every factor, what justification does the workflow give for removing it?
  2. Which two diagnostics (with their reported values) are used to argue that the data are suitable for factor analysis?
  3. After Varimax rotation, what pattern of loadings would confirm that the three-factor structure matches the theoretical dimensions?

Key Points

  1. 1

    Build the initial item pool from both prior scales and qualitative interviews to address gaps in existing measures (e.g., business-focused instruments vs. university-focused concepts).

  2. 2

    Use SPSS EFA diagnostics before interpreting factors: report KMO and Bartlett’s test to justify factor analysis.

  3. 3

    Retain the intended number of factors using eigenvalue criteria (I value > 1) and check cumulative explained variance to support the choice.

  4. 4

    Apply Varimax rotation and enforce a minimum factor loading rule (0.050) to judge whether items represent their intended dimension.

  5. 5

    Remove items that fail to load meaningfully or load onto the wrong factor, re-running EFA after each deletion to stabilize the structure.

  6. 6

    When writing results, include extraction/rotation settings, KMO, Bartlett’s test, commonalities, model-fit residuals, and a clear justification for any item deletions.

  7. 7

    Present the final rotated factor-loading structure in a table with factors labeled according to the three dimensions (ethical, research & development, philanthropic).

Highlights

KMO reached 0.931 and Bartlett’s test was significant, supporting that the item set was suitable for EFA.
A three-factor solution was retained because eigenvalues exceeded one and the three factors explained about 60% of variance.
Two research-and-development items (RD1, RD2) and one philanthropic-related item (PR1) were removed after failing loading criteria, producing a clean factor structure.
The reporting checklist includes extraction method, Varimax rotation, minimum loading threshold, KMO, Bartlett’s test, commonalities, and residual-based model fit.
Item deletion was handled iteratively—removing one problematic item at a time and re-running EFA to confirm the structure improved.

Mentioned

  • EFA
  • SPSS
  • KMO
  • Varimax
  • PR1
  • RD1
  • RD2