Get AI summaries of any video or article — Sign up free
Binary Logistic Regression Analysis using SmartPLS: How to Run, and Interpret the Results. thumbnail

Binary Logistic Regression Analysis using SmartPLS: How to Run, and Interpret the Results.

Research With Fawad·
5 min read

Based on Research With Fawad's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Create a SmartPLS4 logistic regression model by importing data, defining the dependent variable as binary, and coding the target category as 1 and the comparison category as 0.

Briefing

Binary logistic regression in SmartPLS4 can be run as a straightforward “plain regression” model to predict a binary outcome—here, whether respondents choose private versus public sector banks—using a set of predictor variables such as value-added services, perceived cost, and perceived risk. The workflow starts with importing the dataset, creating a new project, defining the dependent variable (Choice) and coding categories so that private bank choice is coded as 1 and public bank choice as 0. After placing the dependent variable and all predictors onto the model canvas, the model is calculated using default settings, producing coefficients, model-fit statistics, classification (confusion) metrics, and significance tests for each predictor.

Interpretation hinges on three layers of output. First, model fit statistics determine whether the predictor-based model improves over a null model with no predictors. Log likelihood and deviance are compared between the null model and the estimated model: higher log likelihood (or, equivalently, a less negative log likelihood) and lower deviance indicate better fit. The session also emphasizes pseudo R-square—specifically NealK r²—as an approximate measure of how much variation in the criterion variable is captured by the predictors. In the example, NealK r² is 0.165, which is translated into an estimated 16.5% change in the criterion variable attributable to the predictors.

Second, information criteria such as AIC and BIC help balance fit against complexity. Lower AIC suggests a better trade-off between model accuracy and overfitting risk, while BIC penalizes complexity more heavily. The reported pattern is that AIC improves from the null model to the estimated model (dropping from 39561 to 37480.5), supporting the idea that the predictors add meaningful explanatory power without excessive cost in complexity. The session also notes that BIC can be higher when many predictors are included, even if other fit measures look good.

Third, classification performance is evaluated using confusion metrics. The model correctly predicts 61 cases when the observed outcome is 0 and correctly predicts 159 cases when the observed outcome is 1. Overall, it gets 64.5% of cases correct, computed as (61 + 159) divided by the total sample size. This “percentage accuracy” is presented as the practical measure of how well the model assigns respondents to the correct bank-choice category.

Finally, the coefficients determine which predictors matter and in what direction. Significance is assessed using p-values (derived from t or z statistics). Odds ratios translate coefficient effects into interpretable changes in the odds of choosing a private bank. An odds ratio greater than 1 means the predictor increases the likelihood of the target event (private bank choice); an odds ratio less than 1 means it decreases that likelihood. In the example, value-added services has an odds ratio of 1.373 and is significant, implying the odds of choosing a private bank are 1.373 times higher when value-added services are present. Perceived cost (odds ratio 0.782, significant) and perceived risk (odds ratio 0.601, significant) both fall below 1, indicating higher perceived cost or risk reduces the odds of choosing private banks relative to public banks.

Taken together, the approach provides a complete decision pipeline: verify model fit, check predictive accuracy, then interpret significant predictors through odds ratios to understand how each factor shifts the probability of private versus public bank choice.

Cornell Notes

Binary logistic regression in SmartPLS4 is used to predict a two-category outcome—here, bank choice—by modeling the odds of choosing private banks (coded as 1) versus public banks (coded as 0). After importing data and assigning the dependent variable and predictors, the model is run to produce coefficients, model-fit statistics (log likelihood, deviance, pseudo R-square), classification performance (confusion metrics), and significance tests. Fit is judged by improvement over a null model: higher log likelihood and lower deviance indicate better fit, while NealK r² provides an approximate explained-variation measure (0.165 in the example). Predictive quality is summarized by overall correct classification (64.5%). Finally, odds ratios interpret direction and magnitude: values above 1 increase the odds of private-bank choice, while values below 1 decrease them.

How does SmartPLS4 represent the binary outcome in logistic regression, and why does coding matter?

The dependent variable must be dichotomous with mutually exclusive categories. In the example, Choice is coded so that choosing private sector banks equals 1 (target group) and choosing public sector banks equals 0 (non-target group). This coding determines what “event” the odds ratio refers to—odds of private-bank choice versus odds of public-bank choice.

What do log likelihood and deviance tell you when comparing the null model to the estimated model?

Log likelihood is the natural log of the likelihood of generating the observed data under the estimated parameters. The session treats a less negative log likelihood (i.e., higher log likelihood) as better fit, showing improvement from the null model (no predictors) to the model with predictors. Deviance measures fit relative to a saturated model; lower deviance indicates better fit. Together, these comparisons indicate whether adding predictors meaningfully improves the model.

Why use pseudo R-square (NealK r²) instead of standard R², and how is it interpreted here?

Logistic regression doesn’t use the same variance-explained logic as linear regression, so pseudo R-square provides an approximate sense of explanatory power rather than a strict “variation explained” metric. The session uses NealK r² as an adjusted pseudo R-square and reports 0.165, interpreted as about 16.5% change in the criterion variable attributable to the predictors (as an approximation).

How is classification accuracy calculated from confusion metrics in SmartPLS4?

Confusion metrics list correct and incorrect predictions by actual category. The example shows 61 correct predictions when the observed outcome is 0 and 159 correct predictions when the observed outcome is 1. Overall accuracy is computed as (61 + 159) divided by the total sample size, yielding 64.5% correct classification.

How do odds ratios translate predictor effects into “odds of choosing private banks”?

Odds ratios convert coefficient effects into interpretable odds changes. An odds ratio greater than 1 means the predictor increases the odds of the target event (private-bank choice). An odds ratio less than 1 means it decreases those odds. In the example, value-added services has an odds ratio of 1.373 (significant), meaning the odds of choosing private banks are 1.373 times higher when value-added services are present. Perceived cost (0.782) and perceived risk (0.601) are significant but below 1, meaning higher cost or risk reduces the odds of choosing private banks relative to public banks.

Review Questions

  1. In logistic regression, what does an odds ratio of exactly 1 imply about the probability of the target event versus the non-target event?
  2. How would you decide whether the model fit is acceptable using log likelihood and deviance when comparing the null model to the estimated model?
  3. If overall correct classification is 64.5%, what does that number represent, and how is it computed from the confusion metrics?

Key Points

  1. 1

    Create a SmartPLS4 logistic regression model by importing data, defining the dependent variable as binary, and coding the target category as 1 and the comparison category as 0.

  2. 2

    Run the model with default settings, then interpret results using three outputs: model fit statistics, classification (confusion) metrics, and predictor coefficients with significance tests.

  3. 3

    Use log likelihood and deviance to judge improvement over the null model: higher log likelihood and lower deviance indicate better fit.

  4. 4

    Report pseudo R-square (NealK r²) as an approximate explained-variation measure rather than a strict R², since logistic regression uses a different underlying logic.

  5. 5

    Evaluate predictive performance with confusion metrics and compute overall accuracy as (correct predictions) divided by total cases.

  6. 6

    Interpret predictor effects through odds ratios: values above 1 increase the odds of the target event; values below 1 decrease those odds.

  7. 7

    Use AIC and BIC to balance fit and complexity, remembering that BIC penalizes complexity more strongly than AIC.

Highlights

NealK r² in the example is 0.165, treated as an approximate 16.5% change in the criterion variable attributable to the predictors.
The model correctly classifies 64.5% of cases, based on 61 correct predictions for outcome 0 and 159 correct predictions for outcome 1.
Value-added services has an odds ratio of 1.373 (significant), meaning it increases the odds of choosing private banks versus public banks.
Perceived cost (odds ratio 0.782) and perceived risk (odds ratio 0.601) are significant predictors that reduce the odds of choosing private banks.

Topics

  • Binary Logistic Regression
  • SmartPLS4 Setup
  • Model Fit Statistics
  • Confusion Metrics
  • Odds Ratio Interpretation

Mentioned