Quick Guide - Part 1 - The Theory of Improving Model Fit in CB-SEM (See Description)
Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Remove indicators with outer loadings below 0.50 and rerun the model after each deletion to see whether fit improves.
Briefing
Improving model fit in CB-SEM starts with a disciplined audit of the measurement model: low outer loadings first, then modification indices, and only after that a targeted look at standardized residual covariances. When fit is poor, the fastest path is to identify weak indicators—specifically loadings below 0.50—remove the problematic indicator, and rerun the model to see whether overall fit improves. This step-by-step “delete and reassess” approach prevents chasing fit problems blindly and keeps attention on which items actually measure the intended construct.
If low loadings don’t resolve the fit issue, the next lever is modification indices, a technique that has sparked debate in the literature but is still widely used in practice to improve fit. The key constraint is theoretical and statistical: modification indices suggest adding covariances, but those covariances must be drawn between error terms of indicators that belong to the same construct (or a similar construct context), not between error terms and latent variables, and not between error terms from different constructs. Every suggested covariance should be checked against theory—shared unexplained variance should make substantive sense, not just reduce chi-square.
After considering modification indices, the process shifts to standardized residual covariances (available in the estimates output). Values above 2 are treated as a warning sign that the model’s implied covariances differ meaningfully from the covariances observed in the data. Those discrepancies can undermine overall fit, and the suggested remedy is usually deletion of the item tied to the problematic residual covariance. However, deletion must be handled carefully: removing too many indicators can damage content validity and distort how the construct was operationalized. The guidance is to avoid over-deleting and to maintain at least three items per construct.
The transcript also offers practical decision rules for modification indices. A covariance suggestion is considered meaningful if the chi-square value improves by at least 10 (not merely a small numerical change). In large models, modification indices can easily exceed 20, and sometimes 30–50, so the focus should be on “extreme” suggestions rather than every minor recommendation. At the same time, multiple high modification indices within a construct raise a red flag: shared unexplained error may reflect method bias from how data were collected, not a true measurement relationship.
Finally, the guidance ties measurement-model troubleshooting back to data quality and questionnaire design. Before modeling, the data should be checked for incorrect or inconsistent values, missing values should be handled via imputation, and the questionnaire should include enough indicators per construct—often at least four to six—because items may later be removed due to low loadings, cross-loadings, or residual problems. The recommended workflow is therefore: clean data, build the model in AMOS, remove low-loading indicators iteratively, apply theoretically justified error covariances suggested by modification indices, address standardized residual covariances above 2 through selective deletion, and only then report the finalized measurement model. The overall message is that fit improvement is not just statistical tinkering; it must stay anchored to theory, content validity, and data collection quality.
Cornell Notes
Poor CB-SEM fit is handled in a staged workflow: start with outer loadings, then use modification indices, and finish with standardized residual covariances. Indicators with loadings below 0.50 are removed, and the model is rerun after each deletion. If fit remains poor, modification indices can guide adding covariances, but only between error terms of indicators within the same construct, and only when theory justifies the shared unexplained variance. After that, standardized residual covariances above 2 signal meaningful gaps between proposed and observed covariances; the usual response is selective item deletion tied to those residuals. Throughout, deletion must protect content validity—keep at least three items per construct and avoid removing too many indicators.
Why begin with outer loadings before touching modification indices?
What covariances are allowed when using modification indices in CB-SEM?
How large should a modification index improvement be to matter?
What do standardized residual covariances above 2 indicate, and what should be done?
How can method bias show up during model-fit improvement?
Review Questions
- What is the recommended order of operations when CB-SEM fit is poor, and what threshold values trigger each step?
- Under what theoretical and statistical conditions is it acceptable to add covariances suggested by modification indices?
- Why is it risky to delete too many indicators, and what minimum number of items per construct is advised?
Key Points
- 1
Remove indicators with outer loadings below 0.50 and rerun the model after each deletion to see whether fit improves.
- 2
Use modification indices only to add theoretically justified covariances between error terms within the same construct, not between error terms and latent variables or across different constructs.
- 3
Treat modification index suggestions as meaningful when they improve chi-square by at least 10; prioritize extreme values in large models rather than every small recommendation.
- 4
Check standardized residual covariances in the estimates output and consider deleting items tied to values above 2, but do so after addressing modification indices.
- 5
Avoid excessive deletion to protect content validity; keep at least three items per construct.
- 6
Improve fit by starting with data quality: correct inconsistent values, handle missing data via imputation, and design questionnaires with enough items per construct (often 4–6) to allow later refinement.