
Online edition (c)2009 Cambridge UP
12.1 Language models 241
questionable whether there is enough training data to do more. Losses from
data sparseness (see the discussion on page
260) tend to outweigh any gains
from richer models. This is an example of the bias-variance tradeoff (cf. Sec-
tion
14.6, page 308): With limited training data, a more constrained model
tends to perform better. In addition, unigram models are more efficient to
estimate and apply than higher-order models. Nevertheless, the importance
of phrase and proximity queries in IR in general suggests that future work
should make use of more sophisticated language models, and some has be-
gun to (see Section
12.5, page 252). Indeed, making this move parallels the
model of van Rijsbergen in Chapter
11 (page 231).
12.1.3 Multinomial distributions over words
Under the unigram language model the order of words is irrelevant, and so
such models are often called “bag of words” models, as discussed in Chap-
ter
6 (page 117). Even though there is no conditioning on preceding context,
this model nevertheless still gives the probability of a particular ordering of
terms. However, any other ordering of this bag of terms will have the same
probability. So, really, we have a multinomial distribution over words. So longMULTINOMIAL
DISTRIBUTION
as we stick to unigram models, the language model name and motivation
could be viewed as historical rather than necessary. We could instead just
refer to the model as a multinomial model. From this perspective, the equa-
tions presented above do not present the multinomial probability of a bag of
words, since they do not sum over all possible orderings of those words, as
is done by the multinomial coefficient (the first term on the right-hand side)
in the standard presentation of a multinomial model:
P(d) =
L
d
!
tf
t
1
,d
!tf
t
2
,d
! ···tf
t
M
,d
!
P(t
1
)
tf
t
1
,d
P(t
2
)
tf
t
2
,d
···P(t
M
)
tf
t
M
,d
(12.7)
Here, L
d
=
∑
1≤i≤M
tf
t
i
,d
is the length of document d, M is the size of the term
vocabulary, and the products are now over the terms in the vocabulary, not
the positions in the document. However, just as with STOP probabilities, in
practice we can also leave out the multinomial coefficient in our calculations,
since, for a particular bag of words, it will be a constant, and so it has no effect
on the likelihood ratio of two different models generating a particular bag of
words. Multinomial distributions also appear in Section
13.2 (page 258).
The fundamental problem in designing language models is that we do not
know what exactly we should use as the model M
d
. However, we do gener-
ally have a sample of text that is representative of that model. This problem
makes a lot of sense in the original, primary uses of language models. For ex-
ample, in speech recognition, we have a training sample of (spoken) text. But
we have to expect that, in the future, users will use different words and in