18. SEMinR Lecture Series - Evaluating Formative Model | Step 3 | Indicator Weights
Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Formative indicator weights are evaluated as the final step after convergent validity and collinearity checks.
Briefing
Formative measurement model evaluation reaches its final checkpoint by testing whether each indicator’s weight is statistically meaningful and whether the indicator truly contributes to forming the construct. In practice, that means running a bootstrap procedure to test the significance and relevance (size) of indicator weights, then using confidence intervals and indicator loadings to decide whether weak or insignificant weights should trigger indicator removal. The stakes are straightforward: formative models can’t rely on internal consistency alone, so indicator-level evidence has to guide retention or deletion.
The process starts after collinearity and convergent validity checks are already completed. For the “indicator weights” step, significance testing depends on bootstrapping. The workflow uses a SEMinR bootstrap function with a SEMinR model specified as the bootstrap target. A large number of bootstrap samples is recommended—10,000 for final reporting—though a smaller run (e.g., 1,000) can be used for preliminary estimation to save computation time. To speed things up, the procedure can automatically detect and use the maximum number of cores available on the machine (via a parallel cores option). A seed can make results reproducible, but it’s suggested to keep defaults unless strict repeatability is required.
After bootstrapping, the bootstrap output is stored in a summary object (via a summary function) that takes the bootstrap results and an alpha level. The transcript uses the default alpha = 0.05 for two-tailed testing, aligning with the convention for indicator weight significance. The resulting output provides indicator weights along with T values (and related statistics). A common rule of thumb is that, for a two-tailed 5% test, T values above 1.96 indicate statistical significance. The transcript also notes other critical thresholds for different alpha levels (e.g., 1% and 10%), but the decision framework in this run centers on 1.96.
In the example output, some formative indicators show significant weights while others fall short of the T-value threshold. That outcome alone doesn’t automatically justify deleting indicators. Instead, confidence intervals offer a second lens: if a weight’s confidence interval does not include zero, the weight is treated as significant and the indicator can be retained; if the interval includes zero, the weight is not statistically significant at the chosen alpha and the indicator becomes a candidate for removal.
When weights are insignificant, the modeler must still check absolute contribution through indicator loadings. The transcript treats loadings of 0.5 or higher as evidence of meaningful absolute contribution. Importantly, even an indicator with a lower weight can still be retained if its loading is strong and its confidence interval excludes zero. In the worked example, indicator weights are mixed—some are insignificant—but the indicator loadings are reported as consistently above 0.5, and the confidence-interval logic supports keeping the indicators. The session ends with the takeaway that indicator retention in formative models depends on both statistical significance (via T values and confidence intervals) and practical contribution (via loadings), not on weight significance alone.
Cornell Notes
Formative model evaluation’s final step tests whether each indicator’s weight is statistically significant and whether the indicator meaningfully contributes to forming the construct. Significance is assessed through bootstrapping (recommended 10,000 samples for final results) and then summarizing bootstrap outputs at a chosen alpha level—here, alpha = 0.05 with two-tailed testing. Indicator weights are judged significant when T values exceed 1.96 and when bias-corrected confidence intervals exclude zero. If weights are insignificant, deletion isn’t automatic; indicator loadings are checked for absolute contribution, with loadings ≥ 0.5 treated as strong evidence to retain indicators. In the example, mixed weight significance still results in retaining all formative indicators because loadings and confidence-interval behavior support their contribution.
Why does formative indicator weight significance require bootstrapping rather than relying on standard parametric tests?
How do T values translate into “significant” indicator weights at alpha = 0.05?
What does it mean if an indicator weight’s confidence interval includes zero?
If an indicator weight is insignificant, why might the indicator still be retained?
What practical decision rule emerges from combining weight significance and loadings?
Review Questions
- What bootstrap settings (sample size, cores, alpha) are used to test formative indicator weights, and why does alpha matter?
- How do you decide significance using both T values and confidence intervals for indicator weights?
- When would you check indicator loadings even if an indicator weight is not significant, and what loading threshold is used as a benchmark?
Key Points
- 1
Formative indicator weights are evaluated as the final step after convergent validity and collinearity checks.
- 2
Significance testing for indicator weights relies on bootstrapping, with 10,000 bootstrap samples recommended for final reporting.
- 3
Use alpha = 0.05 for two-tailed testing; at this level, T values above 1.96 indicate significant indicator weights.
- 4
Confidence intervals guide significance decisions: intervals that exclude zero support retaining an indicator; intervals that include zero suggest non-significance.
- 5
Insignificant weights do not automatically justify deletion; indicator loadings must be checked for absolute contribution.
- 6
Indicator loadings of 0.5 or higher are treated as evidence of meaningful absolute contribution, supporting retention even when some weights are weak.
- 7
A combined decision framework—T values/confidence intervals plus loadings—determines whether formative indicators stay in the measurement model.