Tony E. Smith

NOTES ON LOGISTIC REGRESSION 1. The Logit Model. These notes introduce a special regression model for analyzing dichotomous dependent variables, such as voter opinions (“Yes” vs “No”) or test results (“Pass” vs “Fail”), typically represented by a binary random variable Y 0,1 . To see the need for a special model, note that for any explanatory variables, ( x1 ,.., xk ) , the standard linear model

(1)

Y 0 j 1 j x j , N 0, 2 k

assumes that Y is normally distributed (and hence cannot be binary). A much better way to model this data is to assume that the variables ( x1 ,.., xk ) influence the probability that Y 1 through some appropriate transform of the linear function in (1). The simplest function of this type is the logit function , which assumes in particular that:

(2)

Pr(Y 1 x1 ,.., xk )

exp 0 j 1 j x j k

1 exp 0 j 1 j x j k

and hence that

(3)

Pr(Y 0 x1 ,.., xk ) 1

exp 0 j 1 j x j k

1 exp 0 j 1 j x j k

1

1 exp 0 j 1 j x j k

This is called the (linear) logit model of Y with respect to ( x1 ,.., xk ) .

2. Maximum-Likelihood Estimation of Parameters

If for any given set of data yi , x1i ,.., xki , i 1,.., n it is assumed that the n samples are independent, then the joint probability of the observed values ( y1 ,.., yn ) is simply (4)

Pr( y1 ,.., yn ) i 1 Pr Yi yi x1i ,.., xki n

1

By substituting the appropriate expressions [either (2) or (3)] into each factor of the right hand side of (4), one can write this probability as an explicit function of the unknown parameters 0 , 1 ,.., k . The resulting function is called the likelihood function for

0 , 1 ,.., k given ( y1 ,.., yn ) , and can be written as L 0 , 1 ,.., k y1 ,.., yn

k exp 0 j 1 j x j n i 1 k 1 exp 0 j 1 j x j

yi

1 1 exp k x j 1 j j 0

1 yi

Note that whenever yi 1 , it follows that 1 yi 0 and hence that the term in brackets becomes (2). Conversely, when yi 0 it follows that 1 yi 1 and hence that this term becomes (3). Next observe that by maximizing L with respect to the unknown parameters 0 , 1 ,.., k , one must obtain the parameter values ˆ 0 , ˆ 1 ,.., ˆ k under which the

observed outcomes ( y1 ,.., yn ) are most likely to occur. These parameter values ˆ , ˆ ,.., ˆ are called the maximum-likelihood estimates of , ,.., given

0

1

k

0

1

k

( y1 ,.., yn ) . In the present case this maximum-likelihood estimation procedure is designated as logistic regression, and is available in JMP.

3. An Application using JMP

To illustrate logistic regression in JMP, we employ the weather data set in the JMP file, Rain_Logit.jmp. This contains weather data collected at the same location for each 30 days. Here the dichotomous variable Y is designated as Rained, and has nominal values “Rainy” vs “Dry”. The nominal categorization of this variable is denoted in JMP by the red icon next to Rained in list of columns on the left side of the data window. The two explanatory variables ( x1 , x2 ) correspond to barometric pressure (Pressure) and relative humidity (Humidity) respectively. To run a logistic regression of Rained on (Pressure, Humidity), simply proceed as with any multiple regression by using the Fit Model option, where Y is specified as Rained, and where Pressure and Humidity are added as the explanatory variables. Notice that the Personality box in the upper right corner of the Fit Model window now reads “Nominal Logistic” rather than “Standard Least Squares”. JMP has automatically detected the presence of a nominal variable as Y and selected logistic regression as the default option for this case. When you click Run Model, you will now see that the output results of this maximumlikelihood estimation procedure are substantially different from those of multiple

2

regression. First notice that there are no graphical outputs. (If you run a logistic regression with one explanatory variable using Fit Y by X you will see a graphical output. However, this graph is difficult to interpret, and is best ignored.)

3.1 Parameter Estimation

Start by looking at the Parameter Estimates window, shown below:

Parameter Estimates Term Intercept Pressur Humidity

Estimate -612.89105 21.0482199 -0.0899832

Std Error ChiSquare Prob>ChiSq 299.1737 4.20 0.0405 10.243164 4.22 0.0399 0.0480532 3.51 0.0611

For log odds of Dry/Rainy

Here the Estimate ˆ j and Std Error s j look exactly like those for multiple regression. However, these ˆ j ’s are maximum-likelihood estimates and not least-squares estimates. Moreover, observe that the signs of these beta coefficients appear to be wrong: Rain should be more likely when Pressure decreases and/or Humidity increases. The difficulty here is that JMP has set 1 = “Dry” and 0 = “Rainy”. In making its choice, JMP automatically chooses the first nominal value (alphabetically) of the dependent variable as the event value (Y 1) in computing the probability formulas in (2) and (3) above. In the present case, “Dry” is the event value so that (2) is computed for “Dry”, as can be seen by the fact that the linear formula reported in the columns above is Lin[Dry] rather than Lin[Rainy] . If you wish to ensure that (2) is computed for the event of most interest, you must label this event so that it is first alphabetically. For example, in the file Rain_Logit.jmp, a new variable, Rained_New, has been constructed with nominal values, a_Rainy and b_Dry. If the logistic regression is run on this variable, then JMP will make a_Rainy the event value. Try it. 3.1.1. Interpretation of Beta Values. Unlike linear regression, one cannot interpret the beta values, j , as simple partial derivatives of E (Y | x1 ,.., xk ) with respect to the x j ’s.

However, one can interpret them in terms of “odds ratios”. First, observe that if the odds ratio of Y 1 vs Y 0 for a given set of attributes x1 ,.., xk is denoted by (5)

R ( x1 ,.., xk )

Pr(Y 1 x1 ,.., xk ) Pr(Y 0 x1 ,.., xk )

3

[so that for example, Pr(Y 1 x1 ,.., xk ) .75 implies R ( x1 ,.., xk ) .75 / .25 3 , and hence that odds of the relevant event Y 1 occurring are “3 to 1”], then by (2) and (3) we see that (6)

R ( x1 ,.., xk ) exp 0 j 1 j x j k

Hence if any explanatory variable, x j , is incremented by one unit, then we see that

R( x1 ,.., xi 1,..., xk ) exp 0 j ( x j 1) l j l xl R ( x1 ,.., xi ,..., xk ) exp 0 j x j l xl (7)

l j

exp 0 j ( x j 1) l j l xl 0 j x j l j l xl exp j

Thus if exp j 2 then this would mean that incrementing x j by one unit would

double the odds of the event Y 1 occurring. It is this interpretation of j that is typically adopted. However, the value of “one unit” depends very much on the units of measure. For example, in the present example, a full inch of mercury in barometric pressure is an enormous pressure change. So if we use the dependent variable Rained_New, and take a_Rainy to be the relevant event, then the value j 21.05 yields such a drastic

reduction in the odds ratio [ exp(1 ) exp(21.05) 109 ] that it is difficult to interpret. However, if one redefines the appropriate increment in x j to be of any size (rather

than 1), then the right hand side of (7) becomes exp j . So for example, if it is more meaningful to consider increments in barometric pressure of say a tenth of an inch of mercury, then (7) would yield the value exp(0.1 1 ) exp(2.105) 1/ 8 . For inverse relationships of this type, a more intuitive interpretation might then be to say the model predicts that a fall in the barometric pressure of a tenth of an inch will lead to an eightfold increase in the odds of rain. Note finally that the convention in JMP is to use the range of x j in the given data set as

the relevant “increment” size . In particular, if one right clicks on the top bar (Nominal Logistic Fit) and selects Odds Ratio, then the value observed for Pressure is 0.00000174. To understand this value, use the calculator to find the maximum and minimum pressure values (Formula Statistical Col Minimum and Col Maximum). Here one obtains the values Pressuremax 29.71 and Pressure min 29.08 .

Hence the JMP value is given by exp 29.71 29.08 (21.05) 0.00000174 .

4

3.1.2. Standard Errors and P-Values for Beta Estimates. Next we consider the standard errors, s j , which in this case are now estimates of the asymptotic standard

deviations of the parameter estimates (as the sample size approaches infinity). The ChiSquare value for each is simply the square of the standardized value ˆ j s j [under

the null hypothesis that ˆ j 0 ]. As an illustration, observe that for the Pressure variable,

2

ChiSquare 4.22 (21.05 /10.24) 2 ˆ 1 s1 . The associated Chi-Square random

variable is thus simply the square of a normal variable, with distribution given by Table A.7 in Devore (p.727). In JMP the appropriate P-Value can be calculated more explicitly from the Chi-Square distribution (with one degree of freedom) as follows: (8)

1 ChiSquare Distribution(4.22, 1) 0.0399

Thus the value is reported above as simply Prob>ChiSq = .0399. [Note that is the same as the limiting P-Values for multiple regression, Prob>|z| (i.e., with z replacing t in the limit). In the present case, z ˆ 1 s1 21.05 /10.24 yields the normal P-Value, 2 ( z ) .0399 .] In the present example, these results suggest that both Pressure and Humidity are reasonably good indicators of whether or not it Rained, with Pressure clearly being the more significant of the two. [It should be noted however that there is a serious question as to whether these asymptotic P-Values are appropriate for small samples. Since the usual t-approximation to the distribution of ˆ j s j is not valid for these maximum-likelihood estimates, the convention in most standard software (including JMP) is simply to use the asymptotic normal P-Values for all sample sizes. A discussion of alternative “Quasi t-Tests” can be found in Ben-Akiva, M. and S. Lerman, Discrete Choice Analysis, MIT Press, 1985, p.26.]

3.2 Goodness-of-Fit Measures

Next look at the Whole Model Test window, shown below:

Whole Model Test Model -LogLikelihood Difference 10.339797 Full 7.986132 Reduced 18.325929

DF ChiSquare 2 20.67959

RSquare (U) Observations (or Sum Wgts) Converged by Gradient

5

0.5642 30

Prob>ChiSq ChiSq < .0001.

6

Finally it is of interest to ask whether there is some simple measure of “goodness of fit” in this case which is comparable to R 2 for multiple regression. The short answer is no. However, a number of such measures have been proposed. The one reported in JMP is the so called RSquare (U) value [designated more accurately as the Likelihood Ratio Index (LRI) in Green, W.H., Econometric Analysis, 2nd ed., 1993, p.651]. This value is given (in the present case) by: (14)

LRI 1

ln( L1 ) 7.986 1 .5642 ln( L0 ) 18.326

Notice that since ln( L0 ) ln( L1 ) 0 it follows that 0 ln( L1 ) ln( L0 ) 1 , and hence that LRI has the same value range as R 2 . Moreover, it is also clear that higher values of LRI indicate that the logit model is fitting the data much better than the simple Bernoulli model. But beyond this, there is little that can be said. In particular there is no meaningful notion of “fraction of total variation explained”. An alternative approach, which is much easier to interpret, is simply to look at how well the model actually predicts the dichotomous outcomes. If the estimated value of the probability Pr(Yi 1 xi1 ,.., xik ) for each sample i 1,.., n is denoted by

(15)

ˆ x ,.., x ) P( i1 ik

k exp ˆ 0 j 1 ˆ j xij

k 1 exp ˆ 0 j 1 ˆ j xij

then it is natural to consider the estimated model as “predicting” Yi 1 whenever (16)

ˆ x ,.., x ) P( i1 ik

1 2

If one then defines the outcome prediction of the model for each sample i 1,.., n by

(17)

ˆ x ,.., x ) 1 , P( i1 ik yˆi otherwise 0 ,

1 2

then the prediction success of the model for sample i can be denoted by (18)

1 , Si 0 ,

yi yˆi yi yˆ

With these conventions, a natural measure of goodness of fit for a logistic regression model is given by the prediction success rate

7

(19)

S

1 n

n i 1

Si

This can easily be calculated in JMP as follows: (i)

First, right click on the top bar in the Logistic Regression window, and select Save Probability Formula.

(ii)

This will add four new columns to the data set (Lin[a_Rainy], Prob[b_Dry], Prob[a_Rainy], MostLikely Rained_New). In the present case the two most useful columns are Prob[a_Rainy] and MostLikely Rained_New. The first gives the estimated probability formula in (15) above, and can be used to obtain probability predictions for data values ( x1 ,.., xk ) not in the sample. The second gives precisely the outcome prediction values in (17) above [where in the present case the values are 1 = “a_Rainy” and 0 = “b_Dry”.

(iii)

So to construct the corresponding prediction success, one can add a new column, labeled Success in the file Rain_Logit.jmp, and calculate the success value by the formula for this column, i.e., by Rained _ New MostLikely Rained _ New 1 If else 0

(iv)

Finally, the prediction success rate can be computed as in the column Success Rate by the formula Col Mean Success

(v)

This value shows a 90% success rate, which is quite respectable. (Remember however that this success rate is only for the data used to fit the model, and offers no guarantees for future predictions.)

Now that you are familiar with Logistic Regression, it is important to emphasize you can also do STEPWISE Logistic Regression by simply choosing the option “Stepwise” rather than “Nominal Logistic”. The stepwise procedure here is identical with linear regression, and requires no additional discussion. Finally it should be noted that you can also analyze possible specification errors by letting Rain be an indicator-variable version of Rained with Rain = 1 if and only if Rained = Rainy. Then define appropriate “residuals” by Res = Rain – Prob[Rained]). By plotting each of your explanatory variables against the squared residuals, Res2, you can now look for possible structure in these patterns. For example, if these squared residuals are U-shaped (i.e., largest at each end of the plot) for variable x, then perhaps a quadratic specification of x would yield a better fit.

8