where we introduce the notation 1[] to define a binary outcome. The function 1[] is called
the indicator function,which takes on the value one if the event in brackets is true, and
zero otherwise. Therefore, y is one if y* 0, and y is zero if y* 0. We assume that e is
independent of x and that e either has the standard logistic distribution or the standard nor-
mal distribution. In either case, e is symmetrically distributed about zero, which means that
1 G(z) G(z) for all real numbers z. Economists tend to favor the normality assump-
tion for e,which is why the probit model is more popular than logit in econometrics. In
addition, several specification problems, which we touch on later, are most easily analyzed
using probit because of properties of the normal distribution.
From (17.6) and the assumptions given, we can derive the response probability for y:
P(y 1x) P(y* 0x) P[e (
0
x
)x]
1 G[(
0
x
)] G(
0
x
),
which is exactly the same as (17.2).
In most applications of binary response models, the primary goal is to explain the
effects of the x
j
on the response probability P(y 1x). The latent variable formulation
tends to give the impression that we are primarily interested in the effects of each x
j
on
y*. As we will see, for logit and probit, the direction of the effect of x
j
on E(y*x)
0
x
and on E(yx) P(y 1x) G(
0
x
) is always the same. But the latent
variable y* rarely has a well-defined unit of measurement. (For example, y* might be the
difference in utility levels from two different actions.) Thus, the magnitudes of each
j
are
not, by themselves, especially useful (in contrast to the linear probability model). For most
purposes, we want to estimate the effect of x
j
on the probability of success
P(y 1x), but this is complicated by the nonlinear nature of G().
To find the partial effect of roughly continuous variables on the response probability,
we must rely on calculus. If x
j
is a roughly continuous variable, its partial effect on
p(x) P(y 1x) is obtained from the partial derivative:
g(
0
x
)
j
,whereg(z) (z).
(17.7)
Because G is the cdf of a continuous random variable, g is a probability density function.
In the logit and probit cases, G() is a strictly increasing cdf, and so g(z) 0 for all z.
Therefore, the partial effect of x
j
on p(x) depends on x through the positive quantity g(
0
x
), which means that the partial effect always has the same sign as
j
.
Equation (17.7) shows that the relative effects of any two continuous explanatory vari-
ables do not depend on x: the ratio of the partial effects for x
j
and x
h
is
j
/
h
. In the typical
case that g is a symmetric density about zero, with a unique mode at zero, the largest effect
occurs when
0
x
0. For example, in the probit case with g(z)
(z), g(0)
(0)
1/
2
.40. In the logit case, g(z) exp(z)/[1 exp(z)]
2
, and so g(0) .25.
If, say, x
1
is a binary explanatory variable, then the partial effect from changing x
1
from
zero to one, holding all other variables fixed, is simply
G(
0
1
2
x
2
…
k
x
k
) G(
0
2
x
2
…
k
x
k
).
(17.8)
dG
dz
∂p(x)
∂x
j
Chapter 17 Limited Dependent Variable Models and Sample Selection Corrections 585