14. SPSS AMOS | Factor Loadings in Structural Equation Modelling (SEM)
Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Factor loadings quantify how strongly each indicator reflects its latent construct and are estimated via confirmatory factor analysis in the measurement model.
Briefing
Low factor loadings in structural equation modeling (SEM) can threaten convergent and discriminant validity—but dropping indicators isn’t automatic. Factor loadings measure how strongly each observed indicator reflects its underlying latent construct, and they come from confirmatory factor analysis in the measurement model. When an indicator’s loading is weak, it may contribute little to explaining the construct; when it’s weak but the construct still performs well overall, it might still capture a unique aspect of the concept.
A common rule-of-thumb threshold is around 0.70 for factor loadings, yet the practical decision depends on the construct’s overall reliability and the magnitude of the weakness. If composite reliability and related metrics remain acceptable—such as composite reliability above 0.70 and an AVE (average variance extracted) above 0.50—then low individual loadings do not necessarily invalidate convergent validity. In that situation, an indicator with loading below 0.70 can still be worth keeping, especially when the construct is measured with many items and the majority of indicators load strongly. The weaker item may still help represent a distinct component of the latent construct rather than merely adding noise.
The guidance tightens when loadings fall well below the usual benchmark. If an indicator’s loading is clearly under 0.70 and especially under 0.60, it explains too little variance in that indicator—described as barely accounting for about one-third of the variance. Such an item contributes little to measuring the latent construct and can increase unexplained variance in the model. That imbalance—more unexplained than explained variance—can undermine both convergent validity (indicators reflecting the same construct) and discriminant validity (distinct constructs remaining distinct).
Even when deletion seems justified, the process needs safeguards. Dropping indicators based on one dataset invites criticism that the model is being tuned to chance. To avoid capitalizing on random sampling variation, changes to the measurement model should be verified with additional data. The recommended approach is pretesting or pilot testing: decide in the pilot which items to drop, then collect a final dataset to confirm that the revised factor structure and measurement properties hold. If the same items repeatedly underperform in the second data collection, that pattern supports the deletion as a substantive measurement issue rather than a one-off artifact of the first sample.
Finally, the lecture emphasizes a content-and-reliability lens rather than a purely mechanical one. In social sciences, outer loadings below 0.70 are common, so researchers should examine how removing an indicator affects composite reliability, content validity, and AVE. Items should be considered for removal mainly when deletion improves composite reliability and helps AVE exceed recommended thresholds. In short: low loadings call for diagnostic attention, but indicator removal should be driven by whether it improves overall measurement quality and is stable across samples.
Cornell Notes
Factor loadings in SEM quantify how well each observed indicator reflects its latent construct, derived from confirmatory factor analysis. A loading near or below 0.70 does not automatically require deletion if the construct still meets overall criteria such as composite reliability above 0.70 and AVE above 0.50; weaker items may capture unique components. When loadings drop clearly below 0.60, the indicator explains too little variance and can increase unexplained variance, harming convergent and discriminant validity. Indicator deletion should be validated with a second data collection to avoid capitalizing on chance. Pretesting/pilot testing is recommended: drop items during the pilot, then confirm the revised measurement model in the final dataset.
What exactly does a factor loading represent in SEM, and where does it come from?
If an indicator’s loading is below 0.70, when is it still reasonable to keep it?
At what point does a low loading become a stronger reason to delete an indicator?
Why is a single-sample indicator deletion risky, and what is the recommended fix?
How should pretesting/pilot testing be used when adapting scales to new contexts?
What decision rule should guide deletion beyond the 0.70 loading threshold?
Review Questions
- How do composite reliability and AVE influence the decision to keep an indicator with a loading below 0.70?
- What measurement-quality problem can occur when indicators with loadings below 0.60 are retained?
- What steps help ensure indicator deletion decisions are not driven by chance?
Key Points
- 1
Factor loadings quantify how strongly each indicator reflects its latent construct and are estimated via confirmatory factor analysis in the measurement model.
- 2
A loading near or below 0.70 does not automatically require deletion if composite reliability exceeds 0.70 and AVE exceeds 0.50.
- 3
Indicators with loadings clearly below 0.60 explain too little variance and can increase unexplained variance, weakening convergent and discriminant validity.
- 4
Indicator deletion should be validated with a second data collection to avoid capitalizing on chance from one sample.
- 5
Use pretesting/pilot testing when adapting scales: decide item deletions in the pilot, then confirm the revised measurement model in the final dataset.
- 6
Deletion decisions should consider composite reliability, AVE, and content validity—not just whether a loading falls below 0.70.