9. SEM | SPSS AMOS - Understanding, Assessing, and Improving Model Fit in AMOS
Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Model fit in AMOS is about whether the proposed SEM reproduces the observed covariance structure, not just whether chi-square is significant.
Briefing
Model fit in AMOS is ultimately about whether the proposed measurement and structural relationships reproduce the covariance patterns seen in the data—and the practical workflow is a cycle of diagnosing misfit, making theory-consistent adjustments, and re-running the model until fit indices land in acceptable ranges. After building a model from theory, researchers specify how constructs are measured, collect data, and then use AMOS to estimate parameters and compute overall fit statistics plus standardized parameter estimates. When fit is poor, the fix is not guesswork: it starts with checking indicator quality (outer loadings), then using modification indices and residual diagnostics to target specific sources of discrepancy.
The transcript lays out commonly used fit thresholds for AMOS/SEM. For the chi-square test, the p-value should exceed 0.05, but the guidance is tempered by sensitivity to sample size. Several incremental and absolute fit indices are given: GFI should be above 0.95 and AGFI above 0.90; NFI and NNFI above 0.95; CFI above 0.90 (the session uses a “more liberal” cutoff of 0.90 for CFI and TLI); RMSEA should be below 0.08; and SRMR below 0.08. Convergent validity is also monitored via Average Variance Extracted, with a target above 0.5.
When the model fails these benchmarks, the first diagnostic step is outer loadings: indicators with loadings below 0.50 are candidates for removal. After each deletion, the model must be re-estimated and re-evaluated. If misfit persists, modification indices become the next lever. The key constraint is conceptual and statistical: AMOS suggestions should not be used to correlate error terms across different constructs. Acceptable covariance additions are limited to error terms within the same construct (or otherwise theoretically justified redundancy among indicators), because correlating errors can explain shared unexplained variance without changing the latent structure.
The transcript also emphasizes standardized residual covariances. These values reflect gaps between the model-implied covariance matrix and what the data actually show. Large standardized residual covariances—often flagged when they exceed 2—signal specific indicator relationships that the model fails to reproduce. Deleting items based on residuals can improve fit, but it must be done cautiously to avoid damaging content validity and the operational definition of constructs. A practical process is proposed: clean the dataset (inconsistent values, missingness, imputation), build the model, remove low-loading indicators, apply modification indices only when theoretically defensible, then address residual covariances, and finally report the revised model.
A worked AMOS example demonstrates the loop. Initial fit is weak: CMIN/chi-square is significant, CFI/TLI are below target, and RMSEA is high. The analyst then adds covariances suggested by modification indices (e.g., between specific error terms like E11 and E12), re-runs the model, and observes incremental improvement. Residual diagnostics identify problematic indicators within constructs (notably items labeled SL3, SL6, and FP5). After deleting those indicators, an AMOS “reference point” issue is resolved by fixing a parameter to 1. The final re-run produces substantially improved fit (including CFI/TLI near acceptable levels and SRMR around 0.03), though RMSEA remains comparatively poor. The takeaway is that model improvement is iterative and diagnostic-driven: each adjustment is justified by fit statistics and theory, then validated by re-estimation.
Cornell Notes
Model fit in AMOS measures how well a hypothesized SEM reproduces the covariance structure observed in the data. The workflow starts with theory-based model building, then checks indicator quality (outer loadings), overall fit indices (e.g., CFI/TLI, GFI/AGFI, RMSEA, SRMR), and convergent validity (AVE). If fit is poor, AMOS diagnostics guide changes: remove low-loading indicators, add theoretically justified covariances only among error terms within the same construct, and address large standardized residual covariances (often >2) by considering item deletion. After each change, the model must be re-run and re-evaluated; deleting indicators may require resetting reference parameters (e.g., fixing a parameter to 1).
Why does AMOS model fit matter, and what does “fit” mean in practice?
What are the key fit indices and typical cutoff values mentioned for deciding whether fit is acceptable?
How should low outer loadings be handled when the model fit is poor?
What is the correct way to use modification indices in AMOS?
How do standardized residual covariances guide further model improvement?
What extra step can be required after deleting indicators in AMOS?
Review Questions
- Which fit indices in the transcript are treated as most sensitive to sample size, and how does that affect interpretation?
- When a modification index suggests correlating two error terms, what theory-based constraint determines whether that suggestion is acceptable?
- If a standardized residual covariance exceeds 2 for an indicator, what decision pathway should be followed to improve fit without harming content validity?
Key Points
- 1
Model fit in AMOS is about whether the proposed SEM reproduces the observed covariance structure, not just whether chi-square is significant.
- 2
Use a set of fit indices together (e.g., CFI/TLI, GFI/AGFI, RMSEA, SRMR) and apply the transcript’s cutoffs, including the more liberal CFI/TLI threshold of 0.90.
- 3
Start remediation with outer loadings: remove indicators with standardized loadings below 0.50, then re-run and reassess.
- 4
Apply modification indices only when theoretically justified, and avoid correlating error terms across different constructs.
- 5
Use standardized residual covariances to pinpoint remaining misfit; values above about 2 can justify considering item deletion.
- 6
Protect content validity by not deleting too many items from a construct; the transcript suggests keeping at least ~3 indicators per construct.
- 7
After deleting indicators, check identification/reference points in AMOS; fixing a parameter to 1 may be necessary to re-establish a reference point.