Get AI summaries of any video or article — Sign up free
12. SEMinR Lecture Series - Evaluating Reflective Measurement Model - Step 1: Indicator Reliability thumbnail

12. SEMinR Lecture Series - Evaluating Reflective Measurement Model - Step 1: Indicator Reliability

Research With Fawad·
5 min read

Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Reflective measurement model evaluation in PLS-SEM follows four steps: indicator reliability, internal consistency reliability, convergent validity (AVE), and discriminant validity (HTMT).

Briefing

Evaluating a reflective measurement model in PLS-SEM starts with a practical checklist: confirm indicator reliability, then check internal consistency, followed by convergent validity (via AVE) and discriminant validity (via HTMT). The session frames measurement models as the part of SEM that tests whether constructs are measured reliably and validly, distinct from the structural model that tests relationships among constructs. For reflective models, the quality gates are organized into four steps—indicator reliability, internal consistency reliability, convergent validity, and discriminant validity—each tied to specific statistics and decision rules.

The walkthrough uses a worked example with constructs such as vision development, rewards, collaborative culture, and organizational performance, where all constructs are treated as reflective. It emphasizes that before any reliability or validity statistics can be interpreted, the measurement and structural model must be correctly specified and successfully estimated in the SEMinR workflow. That means loading the SEMinR library, loading the dataset into an object, defining the measurement model by listing constructs and their indicator items, defining the structural paths between constructs, and then running the estimation function. Results are retrieved through a summary object (created via the summary function), and the session stresses that outputs aren’t automatically displayed—you extract them from the stored summary object.

A key operational point is convergence checking. After estimation, the algorithm’s iterations are inspected to ensure the PLS algorithm stopped because the stop criterion was reached—not because it hit the maximum iteration cap. In the example, convergence occurred in five iterations, well below the default maximum of 300, which the session treats as a sign the solution is stable. If convergence fails, two likely causes are highlighted: an overly strict stop criterion (set too small) or data issues such as too small a sample size or indicators with many identical response values that can create singular metrics.

Once the model is estimated and convergence is confirmed, the session moves to indicator reliability. For reflective measurement, indicator reliability is assessed by how much of each indicator’s variance is explained by its construct—computed as the squared indicator loading (because loadings represent the bivariate correlation between an indicator and its construct). The commonly used rule is that squared loadings should exceed 0.50, meaning each indicator explains more than half of its variance through the underlying construct. The session also warns against automatic deletion of indicators that fall below 0.70: researchers should examine whether removing an item actually improves internal consistency reliability and convergent validity enough to justify the change, while also considering content validity. Indicators with extremely low loadings below 0.40 are treated as candidates for elimination, but the decision should still be made carefully.

In the example results, the squared loadings are all above 0.50 (e.g., values around 0.80), so indicator reliability is considered established. The session then signals a transition to construct-level reliability and validity checks as the next step in the reflective measurement model evaluation sequence.

Cornell Notes

Reflective measurement model assessment in PLS-SEM follows a four-step reliability/validity workflow: indicator reliability, internal consistency reliability, convergent validity (using AVE), and discriminant validity (using HTMT). Before interpreting any statistics, the model must be correctly specified (measurement and structural paths), estimated, and shown to converge by checking that the stop criterion is reached within the iteration limit. Indicator reliability is evaluated by squaring each indicator’s loading; squared loadings should exceed 0.50 to indicate that the construct explains more than half the indicator’s variance. Low-loading indicators shouldn’t be removed automatically—item deletion should be justified by meaningful gains in reliability/validity and by preserving content validity. In the example, squared loadings were all above 0.50, so indicator reliability passed.

What distinguishes a measurement model from a structural model in SEM, and why does that matter for reflective assessment?

A measurement model focuses on whether constructs are measured reliably and validly (reliability and validity of indicators/constructs). A structural model focuses on whether relationships among constructs are statistically significant. For reflective measurement assessment, the workflow targets indicator reliability, internal consistency reliability, convergent validity, and discriminant validity—none of which depends on testing path significance.

How is indicator reliability computed for reflective constructs in PLS-SEM?

Indicator reliability is assessed by the variance explained by each indicator’s construct. In practice, it is computed as the squared indicator loading, where the loading is the bivariate correlation between an indicator and its construct. The rule used is that squared loadings should be greater than 0.50 (meaning the construct explains more than half of the indicator’s variance).

What should researchers do if some indicator loadings are below 0.70?

The session discourages automatic deletion solely because loadings fall under 0.70. Instead, researchers should examine whether removing the indicator improves internal consistency reliability and convergent validity enough to cross the required thresholds. Content validity also matters: if deletion harms the construct’s conceptual coverage, the indicator may be retained even with weaker loadings. Indicators with very low loadings below 0.40 are treated as more clear-cut candidates for elimination.

Why check algorithm convergence before interpreting reliability and validity results?

Convergence indicates the PLS algorithm found a stable solution. The session recommends inspecting the iterations in the summary output to confirm the stop criterion was reached before the maximum iteration limit (default 300). If convergence fails, likely causes include an overly small stop criterion or data problems such as too small a sample size or indicators with many identical values that can produce singular metrics.

How are results accessed after estimating a PLS-SEM model in the SEMinR workflow described here?

After running the estimation function, the output is stored in an object (e.g., a model object). A summary function is then called to create a summary object (e.g., summary_simple). Reliability/validity statistics aren’t automatically printed; they are extracted from the summary object using the dollar sign and the relevant sub-object (for example, iterations or loadings/indicator reliability).

Review Questions

  1. What statistic and threshold are used to judge indicator reliability in reflective measurement models, and how is it calculated from loadings?
  2. If an indicator has a loading below 0.70, what decision logic should be used before deleting it?
  3. What two main reasons are given for PLS-SEM failing to converge within the maximum iterations?

Key Points

  1. 1

    Reflective measurement model evaluation in PLS-SEM follows four steps: indicator reliability, internal consistency reliability, convergent validity (AVE), and discriminant validity (HTMT).

  2. 2

    Correctly specify the measurement model (constructs and their reflective indicators) and the structural paths before running estimation; otherwise errors will prevent valid results.

  3. 3

    Check algorithm convergence by verifying the stop criterion is reached within the iteration limit (default maximum 300) rather than relying on the model running without errors.

  4. 4

    Indicator reliability is assessed using squared loadings; values above 0.50 indicate the construct explains more than half the indicator’s variance.

  5. 5

    Do not delete indicators automatically when loadings fall below 0.70; deletion should be justified by improvements in reliability/validity and by preserving content validity.

  6. 6

    Indicators with very low loadings below 0.40 are treated as strong candidates for elimination, but the decision should still consider measurement coverage.

  7. 7

    Use the summary object to extract loadings, indicator reliability, and convergence diagnostics; results are not automatically displayed.

Highlights

Indicator reliability for reflective models is computed as the squared loading, with a common benchmark of >0.50.
Convergence should be verified by checking iterations—stable solutions stop due to the stop criterion well before the maximum iteration cap.
Low-loading indicators shouldn’t be removed just because they miss 0.70; item deletion must be justified by gains in reliability/validity and by content validity.
The workflow relies on a stored summary object from which loadings, iterations, and reliability/validity statistics are extracted.

Topics

Mentioned

  • PLS-SEM
  • HTMT
  • AVE
  • IV
  • SEM
  • PLS
  • CC