Rather than having to personally determine the influence of certain observations, it is
sometimes useful to have statistics that can detect such influential observations. These sta-
tistics do exist, but they are beyond the scope of this text. (See, for example, Belsley, Kuh,
and Welsch [1980].)
Before ending this section, we mention another approach to dealing with influen-
tial observations. Rather than trying to find outlying observations in the data before
applying least squares, we can use an estimation method that is less sensitive to out-
liers than OLS. This obviates the need to explicitly search for outliers before or dur-
ing estimation. One such method, which is becoming more and more popular among
applied econometricians, is called least absolute deviations (LAD). The LAD esti-
mator minimizes the sum of the absolute deviations of the residuals, rather than the
sum of squared residuals. It is known that LAD is designed to estimate the effects of
explanatory variables on the conditional median,rather than the conditional mean, of
the dependent variable. Because the median is not affected by large changes in extreme
observations, the parameter estimates obtained by LAD are resilient to outlying obser-
vations. (See Section A.1 for a brief discussion of the sample median.) In choosing the
estimates, OLS attaches much more importance to large residuals because each resid-
ual gets squared.
Although LAD helps to guard against outliers, it does have some drawbacks. First,
there are no formulas for the estimators; they can only be found by using iterative meth-
ods on a computer. A related issue is that obtaining standard errors of the estimates is
somewhat more complicated than obtaining the standard errors of the OLS estimates.
These days, with such powerful computers, concerns of this type are not very important,
unless LAD is applied to very large data sets with many explanatory variables. A second
drawback, at least in smaller samples, is that all statistical inference involving LAD esti-
mators is justified only asymptotically. With OLS, we know that, under the classical lin-
ear model assumptions, t statistics have exact t distributions, and F statistics have exact F
distributions. While asymptotic versions of these tests are available for LAD, they are jus-
tified only in large samples.
A more subtle but important drawback to LAD is that it does not always consistently
estimate the parameters appearing in the conditional mean function, E(yx
1
,...,x
k
). As
mentioned earlier, LAD is intended to estimate the effects on the conditional median.
Generally, the mean and median are the same only when the distribution of y given the
covariates x
1
,...,x
k
is symmetric about
0
1
x
1
...
k
x
k
. (Equivalently, the popula-
tion error term, u, is symmetric about zero.) Recall that OLS produces unbiased and
consistent estimators of the parameters in the conditional mean whether or not the error
distribution is symmetric; symmetry does not appear among the Gauss-Markov assump-
tions. When LAD and OLS are applied to cases with asymmetric distributions, the
estimated partial effect of, say, x
1
, obtained from LAD can be very different from the
partial effect obtained from OLS. But such a difference could just reflect the difference
between the median and the mean and might not have anything to do with outliers. See
Computer Exercise C9.9 for an example.
If we assume that the population error u in model (9.2) is independent of (x
1
,...,x
k
),
then the OLS and LAD slope estimates should differ only by sampling error whether or
not the distribution of u is symmetric. The intercept estimates generally will be different
332 Part 1 Regression Analysis with Cross-Sectional Data