CHAPTER 8 
 
TEACHING NOTES 
 
This is a good place to remind students that homoskedasticity played no role in showing that 
OLS is unbiased for the parameters in the regression equation.  In addition, you should probably  
mention that there is nothing wrong with the R-squared or adjusted R-squared as goodness-of-fit 
measures.  The key is that these are estimates of the population R-squared, 1 – [Var(u)/Var(y)], 
where the variances are the unconditional variances in the population.  The usual R-squared, and 
the adjusted version, consistently estimate the population R-squared whether or not Var(u|x) = 
Var(y|x) depends on x.  Of course, heteroskedasticity causes the usual standard errors, t statistics, 
and F statistics to be invalid, even in large samples, with or without normality. 
 
By explicitly stating the homoskedasticity assumption as conditional on the explanatory 
variables that appear in the conditional mean, it is clear that only heteroskedasticity that depends 
on the explanatory variables in the model affects the validity of standard errors and test statistics.  
This is why the Breusch-Pagan test, as I have presented it, and the White test, are ideally suited 
for testing for relevant forms of heteroskedasticity.  If heteroskedasticity depends on an 
exogenous variable that does not also appear in the mean equation, this can be exploited in 
weighted least squares for efficiency, but only rarely is such a variable available.  One case 
where such a variable is available is when an individual-level equation has been aggregated.  I 
discuss this case in the text but I rarely have time to teach it. 
 
As I mention in the text, other traditional tests for heteroskedasticity, such as the Park and 
Glejser tests, do not directly test what we want, or are too restrictive.  The Goldfeld-Quandt test 
only works when there is a natural way to order the data based on one independent variable.  
This is rare in practice, especially for cross-sectional applications. 
 
Some argue that weighted least squares is a relic, and is no longer necessary given the 
availability of heteroskedasticity-robust standard errors and test statistics.  While I am somewhat 
sympathetic to this argument, it presumes that we do not care much about efficiency.  Even in 
large samples, the OLS estimates may not be precise enough to learn much about the population 
parameters.  With substantial heteroskedasticity, we might do better with weighted least squares, 
even if the weighting function is misspecified.  As mentioned in Question 8.4 on page 280, one 
can (and perhaps should) compute robust standard errors after weighted least squares.  These 
would be directly comparable to the heteroskedasiticity-robust standard errors for OLS. 
 
Weighted least squares estimation of the LPM is a nice example of feasible GLS, at least when 
all fitted values are in the unit interval.  Interestingly, in the LPM examples and exercises, the 
heteroskedasticity-robust standard errors often differ by only small amounts from the usual 
standard errors.  However, in a couple of cases the differences are notable, as in Computer 
Exercise 8.12.
 
62