
260 LEAST SQUARES, LINEAR MODELS AND BEYOND
Inserting this in the formula above, we find that the estimate of θ is given by
ˆ
θ = ˆm
2
−
12
ˆm
1
/
11
. This is what we obtained in Section 8.8.
The two-step method is most useful in cases where there is a nonlinear relationship to be
described. From the discussion above we know that the equation to solve in order to estimate
θ for the model m = f (θ) is the GLS equation
f
(θ)
t
−1
(ˆm − f (θ)) = 0.
In order to obtain knowledge about θ we can either compute the variance of the stochastic
variable on the left here, or use the fact that the original assumptions mean that
Q(θ) = (ˆm − f (θ))
t
−1
(ˆm − f (θ)) ∈ χ
2
(p).
This gives us a confidence function to work with. As above, there are modifications in the
detail, depending on what information we have on . If we only have an estimate of
we need to modify the distribution appropriately, or appeal to large-sample theory and use
the estimated one as fixed to obtain reasonable approximations. Here ˆm has been presented
as adjusted means from a first-step analysis, but what it really consists of are estimates of
parameters from the model of this first analysis.
So far we have discussed how to get simultaneous confidence in all the model parameters.
What if we want a simultaneous confidence region for only a subset of these, with the remaining
ones considered nuisance parameters? More generally, what if we want a confidence interval
for a particular combination of the model parameters, and not a complex confidence region
for them all? This question was discussed, albeit in a special case, in Section 7.7, and a very
brief and superficial description of the method used, profiling, is as follows. Suppose we
want confidence information for a parameter η = g(θ).
1
Instead of minimizing Q(θ) to get an
estimate of θ, we then minimize it under constraints g(θ) = η for different η. This gives us a
function Q(η) of the new parameter alone, and we can obtain univariate confidence statements
about this by using the confidence function χ
1
(Q(η)). This is a general method that produces
reasonable results on most occasions, at least if the sample is not too small.
9.6 The effect of missing covariates
The classical ANOVA model and the logistic model, together with the other models for
binomial data discussed above, are members of a larger family of regression models, called
generalized linear models (GLMs). One common property of these is that the mean of
the outcome variable is modeled in terms of covariates, as an expression of the form
E(Y |X = x) = h(xβ), for some function h(u) called a response function. An important ques-
tion which then arises is the following. Suppose that there is a true model containing one set
of covariates, but that we have omitted to measure some of these and model the mean using
only the ones we have measured. What are the consequences? In the ANOVA situation we get
more unexplained residual variance, but no other effect on what we estimate. In the general
case, including the logistic model, the situation is more complicated.
1
This does not have to be one parameter; the function g(θ) can define more than one parameter, but it is simpler
if we think of it as a single parameter.