Pseudo-R-squared
In statistics, pseudo-R-squared values are used when the outcome variable is nominal or ordinal such that the coefficient of determination R2 cannot be applied as a measure for goodness of fit and when a likelihood function is used to fit a model. In linear regression, the squared multiple correlation, R2 is used to assess goodness of fit as it represents the proportion of variance in the criterion that is explained by the predictors. In logistic regression analysis, there is no agreed upon analogous measure, but there are several competing measures each with limitations. Some commonly used indices are examined in this article: Likelihood ratio R2L Unadjusted and adjusted geometric mean squared improvement R2M and R2N Tjur R2T
In statistics, pseudo-R-squared values are used when the outcome variable is nominal or ordinal such that the coefficient of determination R2 cannot be applied as a measure for goodness of fit and when a likelihood function is used to fit a model.
In linear regression, the squared multiple correlation, R2 is used to assess goodness of fit as it represents the proportion of variance in the criterion that is explained by the predictors.[1] In logistic regression analysis, there is no agreed upon analogous measure, but there are several competing measures each with limitations.[1][2]
Some commonly used indices are examined in this article:
- Likelihood ratio R2L
- Unadjusted and adjusted geometric mean squared improvement R2M and R2N
- Tjur R2T
R2L by McFadden
[edit]The pseudo R2 credited to McFadden (sometimes called likelihood ratio index[3] or log-likelihood ratio R2[4]) is defined as[5]
or expressed using deviance[1] rather than likelihood
and is preferred over R2M by Paul D. Allison.[2] The two expressions R2L and R2M are then related respectively by,
This is the most analogous index to the squared multiple correlations in linear regression.[6] It represents the proportional reduction in the deviance wherein the deviance is treated as a measure of variation analogous but not identical to the variance in linear regression analysis.[6] One limitation of the likelihood ratio R2 is that it is not monotonically related to the odds ratio,[1] meaning that it does not necessarily increase as the odds ratio increases and does not necessarily decrease as the odds ratio decreases.
Adjusted
[edit]The adjusted version includes the number of predictors K as a penalisation term[7]
R2M by Cox and Snell
[edit]R2M is an alternative index of goodness of fit described by Maddala in 1983 attributed to Cox and Snell[4] related to the R2 value from linear regression (sometimes called geometric mean squared improvement per observation R2[4]).[2] It is given by:
where LM and L0 are the likelihoods for the model being fitted and the null model, respectively, \ln denotes the natural logarithm, and n is the number of observations. L0 is the model with just an intercept.[4] The Cox and Snell index corresponds to the standard R2 in case of a linear model with normal error. In certain situations, R2M may be problematic as its maximum value is which means it never takes the value of one. For example, for logistic regression, the upper bound is for a symmetric marginal distribution of events and decreases further for an asymmetric distribution of events.[2]
Adjusted R2M by Nagelkerke
[edit]Nico Nagelkerke in a highly cited Biometrika paper,[8] provides a correction to the Cox and Snell R2 so that the maximum value is equal to 1. This correction is done by dividing RM by its upper bound.[2][4]
Nevertheless, the Cox and Snell and likelihood ratio R2s show greater agreement with each other than either does with the adjusted Nagelkerke R2.[1] Of course, this might not be the case for values exceeding 0.75 as the Cox and Snell index is capped at this value[citation needed]. The likelihood ratio R2 is often preferred[citation needed] to the alternatives as it is most analogous to R2 in linear regression, is independent of the base rate (both Cox and Snell and Nagelkerke R2s increase as the proportion of cases increase from 0 to 0.5) and varies between 0 and 1.
R2T by Tjur
[edit]Tjur proposed an alternative measure of R2T,[9] which is in two steps:
- For each level of the dependent variable, find the mean of the predicted probabilities of an event.
- Take the absolute value of the difference between these means
This quantity is not an per se, however Tjur shows how it is related to usual metrics.
Example
[edit]This displays R output from calculating pseudo-r-squared values using the "pscl" package by Simon Jackman. The pseudo-R-squared calculated using the formule named for McFadden is labelled “McFadden”. Next to this, the pseudo-r-squared by Cox and Snell is labelled “r2ML” and this type of pseudo-R-squared By Cox and Snell is sometimes simply called “ML”. The last value listed, labelled “r2CU” is the pseudo-r-squared by Nagelkerke and is the same as the pseudo-r-squared by Cragg and Uhler.
Interpretation
[edit]A word of caution is in order when interpreting pseudo-R2 statistics. The reason these indices of fit are referred to as pseudo R2 is that they do not represent the proportionate reduction in error as the R2 in linear regression does.[1] Linear regression assumes homoscedasticity, that the error variance is the same for all values of the criterion. Logistic regression will always be heteroscedastic – the error variances differ for each value of the predicted score. For each value of the predicted score there would be a different value of the proportionate reduction in error. Therefore, it is inappropriate to think of R2 as a proportionate reduction in error in a universal sense in logistic regression.[1]
See also
[edit]References
[edit]- ^ a b c d e f g Cohen, Jacob; Cohen, Patricia; West, Steven G.; Aiken, Leona S. (2002). Applied Multiple Regression/Correlation Analysis for the Behavioral Sciences (3rd ed.). Routledge. p. 502. ISBN 978-0-8058-2223-6.
- ^ a b c d e Allison, Paul D. "Measures of fit for logistic regression" (PDF). Statistical Horizons LLC and the University of Pennsylvania.
- ^ Hardin, J. W., Hilbe, J. M. (2007). Generalized linear models and extensions. USA: Taylor & Francis. Page 60, Google Books
- ^ a b c d e Menard, Scott (2000). "Coefficients of Determination for Multiple Logistic Regression". American Statistician. 54 (1): 17–24. doi:10.2307/2685605.
- ^ McFadden, Daniel (1972). "Conditional logit analysis of qualitative choice behaviour". Working Paper Np. 199/BART 10: 23.
- ^ a b Menard, Scott W. (2002). Applied Logistic Regression (2nd ed.). SAGE. ISBN 978-0-7619-2208-7. [page needed]
- ^ "FAQ: What are pseudo R-squareds?". UCLA. Retrieved 2026-03-25.
- ^ Nagelkerke, N. J. D. (1991). A Note on a General Definition of the Coefficient of Determination. Biometrika, 78(3), 691–692. https://doi.org/10.2307/2337038
- ^ Tjur, Tue (2009). "Coefficients of determination in logistic regression models". American Statistician. 63 (4): 366–372. doi:10.1198/tast.2009.08210. S2CID 121927418.