Get AI summaries of any video or article — Sign up free
9. SEM | SPSS AMOS - Understanding, Assessing, and Improving Model Fit in AMOS thumbnail

9. SEM | SPSS AMOS - Understanding, Assessing, and Improving Model Fit in AMOS

Research With Fawad·
5 min read

Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Model fit in AMOS is about whether the proposed SEM reproduces the observed covariance structure, not just whether chi-square is significant.

Briefing

Model fit in AMOS is ultimately about whether the proposed measurement and structural relationships reproduce the covariance patterns seen in the data—and the practical workflow is a cycle of diagnosing misfit, making theory-consistent adjustments, and re-running the model until fit indices land in acceptable ranges. After building a model from theory, researchers specify how constructs are measured, collect data, and then use AMOS to estimate parameters and compute overall fit statistics plus standardized parameter estimates. When fit is poor, the fix is not guesswork: it starts with checking indicator quality (outer loadings), then using modification indices and residual diagnostics to target specific sources of discrepancy.

The transcript lays out commonly used fit thresholds for AMOS/SEM. For the chi-square test, the p-value should exceed 0.05, but the guidance is tempered by sensitivity to sample size. Several incremental and absolute fit indices are given: GFI should be above 0.95 and AGFI above 0.90; NFI and NNFI above 0.95; CFI above 0.90 (the session uses a “more liberal” cutoff of 0.90 for CFI and TLI); RMSEA should be below 0.08; and SRMR below 0.08. Convergent validity is also monitored via Average Variance Extracted, with a target above 0.5.

When the model fails these benchmarks, the first diagnostic step is outer loadings: indicators with loadings below 0.50 are candidates for removal. After each deletion, the model must be re-estimated and re-evaluated. If misfit persists, modification indices become the next lever. The key constraint is conceptual and statistical: AMOS suggestions should not be used to correlate error terms across different constructs. Acceptable covariance additions are limited to error terms within the same construct (or otherwise theoretically justified redundancy among indicators), because correlating errors can explain shared unexplained variance without changing the latent structure.

The transcript also emphasizes standardized residual covariances. These values reflect gaps between the model-implied covariance matrix and what the data actually show. Large standardized residual covariances—often flagged when they exceed 2—signal specific indicator relationships that the model fails to reproduce. Deleting items based on residuals can improve fit, but it must be done cautiously to avoid damaging content validity and the operational definition of constructs. A practical process is proposed: clean the dataset (inconsistent values, missingness, imputation), build the model, remove low-loading indicators, apply modification indices only when theoretically defensible, then address residual covariances, and finally report the revised model.

A worked AMOS example demonstrates the loop. Initial fit is weak: CMIN/chi-square is significant, CFI/TLI are below target, and RMSEA is high. The analyst then adds covariances suggested by modification indices (e.g., between specific error terms like E11 and E12), re-runs the model, and observes incremental improvement. Residual diagnostics identify problematic indicators within constructs (notably items labeled SL3, SL6, and FP5). After deleting those indicators, an AMOS “reference point” issue is resolved by fixing a parameter to 1. The final re-run produces substantially improved fit (including CFI/TLI near acceptable levels and SRMR around 0.03), though RMSEA remains comparatively poor. The takeaway is that model improvement is iterative and diagnostic-driven: each adjustment is justified by fit statistics and theory, then validated by re-estimation.

Cornell Notes

Model fit in AMOS measures how well a hypothesized SEM reproduces the covariance structure observed in the data. The workflow starts with theory-based model building, then checks indicator quality (outer loadings), overall fit indices (e.g., CFI/TLI, GFI/AGFI, RMSEA, SRMR), and convergent validity (AVE). If fit is poor, AMOS diagnostics guide changes: remove low-loading indicators, add theoretically justified covariances only among error terms within the same construct, and address large standardized residual covariances (often >2) by considering item deletion. After each change, the model must be re-run and re-evaluated; deleting indicators may require resetting reference parameters (e.g., fixing a parameter to 1).

Why does AMOS model fit matter, and what does “fit” mean in practice?

Fit is the feasibility of the proposed model with the data—whether the data’s covariance patterns are consistent with the specified measurement/structural relationships. AMOS estimates parameters and reports overall model fit statistics plus parameter estimates; poor fit means the model-implied covariance matrix diverges from the observed covariance matrix.

What are the key fit indices and typical cutoff values mentioned for deciding whether fit is acceptable?

The transcript lists common thresholds: chi-square p-value > 0.05 (noting sensitivity to sample size), GFI > 0.95 and AGFI > 0.90, NFI and NNFI > 0.95, CFI and TLI using a more liberal cutoff of 0.90, RMSEA < 0.08, SRMR < 0.08, and AVE > 0.5 for convergent validity.

How should low outer loadings be handled when the model fit is poor?

Indicators with standardized regression weights (outer loadings) below 0.50 are flagged for potential deletion. The process is stepwise: remove a low-loading indicator, re-run the model, and reassess fit indices and residual diagnostics before making further changes.

What is the correct way to use modification indices in AMOS?

Modification indices suggest adding covariances, but the transcript stresses a strict rule: do not correlate error terms across different constructs. Correlating error terms is acceptable when they belong to a similar construct and is justified by theory (e.g., redundancy among indicators within a construct). The analyst also uses the magnitude of change as a guide (e.g., Ki-square change not improving enough by at least ~10 is less compelling), and watches for method bias when many high modification suggestions appear.

How do standardized residual covariances guide further model improvement?

Standardized residual covariances indicate standardized differences between model-implied covariances and observed covariances. Values above about 2 are treated as problematic, pointing to specific indicator relationships the model fails to capture. The transcript recommends addressing modification indices first, then using residuals to decide whether to delete items—while protecting content validity (e.g., keeping at least ~3 items per construct).

What extra step can be required after deleting indicators in AMOS?

Deleting an indicator can remove a reference point needed for identification. In the example, after removing indicators, AMOS required fixing a parameter to 1 (via the Parameters section) to restore a reference point before the model could be properly estimated.

Review Questions

  1. Which fit indices in the transcript are treated as most sensitive to sample size, and how does that affect interpretation?
  2. When a modification index suggests correlating two error terms, what theory-based constraint determines whether that suggestion is acceptable?
  3. If a standardized residual covariance exceeds 2 for an indicator, what decision pathway should be followed to improve fit without harming content validity?

Key Points

  1. 1

    Model fit in AMOS is about whether the proposed SEM reproduces the observed covariance structure, not just whether chi-square is significant.

  2. 2

    Use a set of fit indices together (e.g., CFI/TLI, GFI/AGFI, RMSEA, SRMR) and apply the transcript’s cutoffs, including the more liberal CFI/TLI threshold of 0.90.

  3. 3

    Start remediation with outer loadings: remove indicators with standardized loadings below 0.50, then re-run and reassess.

  4. 4

    Apply modification indices only when theoretically justified, and avoid correlating error terms across different constructs.

  5. 5

    Use standardized residual covariances to pinpoint remaining misfit; values above about 2 can justify considering item deletion.

  6. 6

    Protect content validity by not deleting too many items from a construct; the transcript suggests keeping at least ~3 indicators per construct.

  7. 7

    After deleting indicators, check identification/reference points in AMOS; fixing a parameter to 1 may be necessary to re-establish a reference point.

Highlights

Fit indices are treated as a diagnostic dashboard: CFI/TLI, GFI/AGFI, RMSEA, SRMR, and AVE each signal different kinds of misfit or validity problems.
Modification indices are constrained by theory: correlating error terms is acceptable only within the same construct (or similarly justified redundancy), not across constructs.
Standardized residual covariances act like a map of where the model’s implied relationships diverge from the data; values above ~2 are treated as red flags.
In the worked example, deleting problematic indicators (SL3, SL6, FP5) plus resetting a reference parameter to 1 leads to a major improvement, including SRMR around 0.03.

Topics

Mentioned

  • Fawad
  • IBM SPSS
  • AMOS
  • SEM
  • GFI
  • AGFI
  • NFI
  • NNFI
  • CFI
  • TLI
  • RMSEA
  • SRMR
  • AVE
  • CMIN
  • Ki Square
  • RMR