13. SEMinR Series - Reflective Model | Reliability and Convergent Validity
Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Composite reliability (rho C), Cronbach’s alpha, and rho A are used to assess internal consistency reliability in reflective measurement models.
Briefing
Reflective measurement model checks hinge on three linked quality tests: internal consistency reliability, convergent validity, and (later) discriminant validity. In this step, reliability is assessed first using composite reliability (often labeled “rho C”), with Cronbach’s alpha as a more conservative alternative and “rho A” as a middle-ground coefficient. Composite reliability values above 0.70 are typically treated as reliable; 0.60–0.70 is acceptable in exploratory work; and 0.70–0.90 is generally satisfactory to good. Values above 0.90—especially above 0.95—raise a red flag: they can indicate redundant indicators, which can reduce construct validity. Extremely high reliability may also reflect undesirable response patterns such as straight-lining, where respondents choose the same response option across items, inflating correlations among indicator error terms.
Because Cronbach’s alpha and composite reliability rely on different assumptions, the reliability picture can shift. Cronbach’s alpha assumes “tau-equivalence,” meaning all indicator loadings are equal in the population; when that assumption fails, alpha tends to produce lower reliability estimates. Composite reliability (rho C) can be too liberal under the same circumstances. To balance these extremes, rho A is introduced as a coefficient that usually falls between Cronbach’s alpha and composite reliability. A practical rule of thumb used here is that rho C, rho A, and Cronbach’s alpha should exceed 0.70 (with rho A and alpha often expected to exceed 0.50), and rho A should sit between alpha and rho C. In the worked example, the reliability outputs for multiple constructs clear the 0.70 threshold across the three measures, indicating the indicators within each construct are internally consistent.
After reliability, convergent validity checks whether items intended to measure the same construct actually converge on that underlying concept. Convergent validity is quantified using Average Variance Extracted (AVE), computed as the grand mean of squared indicator loadings: square each loading, sum them, then divide by the number of indicators. AVE is interpreted as the proportion of variance in the indicators captured by the construct; the minimum acceptable level is 0.50. In the example, all constructs show AVE values above 0.50, so the indicators are judged to represent their intended constructs well enough to establish convergent validity.
Finally, there’s guidance on reporting results: present the measurement model, then report factor (indicator) loadings, construct reliability (including composite reliability and the reliability coefficients), and AVE for convergent validity. The transcript also notes how to extract these metrics in R (via a reliability summary object) and how to visualize reliability with a plot that includes Cronbach’s alpha, rho A, and rho C. With reliability and convergent validity established, the next phase moves to discriminant validity for reflective measurement models.
Cornell Notes
Internal consistency reliability for reflective constructs is evaluated using three coefficients: Cronbach’s alpha (conservative), composite reliability rho C (often liberal), and rho A (a middle coefficient). Threshold guidance used here treats values above 0.70 as reliable, with 0.60–0.70 acceptable in exploratory research; values above 0.90 (and especially above 0.95) can suggest redundancy among indicators and potentially inflated correlations from straight-lining. Convergent validity is then assessed with Average Variance Extracted (AVE), calculated from squared indicator loadings; AVE reflects how much variance in indicators the construct explains. An AVE of 0.50 or higher indicates that indicators converge on the intended construct. In the example, reliability clears the 0.70 threshold and all constructs have AVE above 0.50, so both reliability and convergent validity are established.
Why can very high reliability (e.g., above 0.95) be a problem in reflective measurement models?
How do Cronbach’s alpha, rho C, and rho A differ, and why does that matter?
What are the computation steps for Average Variance Extracted (AVE)?
What threshold is used for convergent validity, and what does it mean?
What does it mean if reliability measures exceed 0.70 but AVE is below 0.50?
Review Questions
- What reliability thresholds are used for composite reliability (rho C), Cronbach’s alpha, and rho A, and what does a value above 0.95 imply?
- How is AVE calculated from indicator loadings, and why does AVE ≥ 0.50 support convergent validity?
- Why is rho A considered a compromise between Cronbach’s alpha and composite reliability (rho C)?
Key Points
- 1
Composite reliability (rho C), Cronbach’s alpha, and rho A are used to assess internal consistency reliability in reflective measurement models.
- 2
Reliability above 0.70 is generally treated as reliable; 0.60–0.70 is acceptable for exploratory research; values above 0.90 (especially above 0.95) can indicate redundancy and reduced construct validity.
- 3
Cronbach’s alpha is conservative because it assumes tau-equivalence (equal indicator loadings), which can lower reliability estimates when loadings differ.
- 4
rho C can be too liberal under non-equal loadings, so rho A is used as a middle coefficient that typically falls between alpha and rho C.
- 5
Convergent validity is assessed with AVE, computed from squared indicator loadings; AVE ≥ 0.50 indicates the construct explains at least half of indicator variance.
- 6
Reporting should include factor loadings, construct reliability coefficients, and AVE for each construct before moving to discriminant validity.