
you are here 4 233
using discrete probability distributions
Q:
So if X and Y are games, does
aX + bY mean a games of X and b games
of Y?
A: aX + bY actually refers to two linear
transforms added together. In other words,
the underlying values of X and Y are
changed. This is different from independent
observations, where each game would be an
independent observation.
Q:
I can’t see when I’d ever want to
use X – Y. Does it have a purpose?
A: X – Y is really useful if you want to find
the difference between two variables.
E(X – Y) is a bit like saying “What do you
expect the difference between X and Y to
be”, and Var(X – Y) tells you the variance.
Q:
Why do you add the variances for
X – Y? Surely you’d subtract them?
A: At first it sounds counterintuitive,
but when you subtract one variable from
another, you actually increase the amount
of variability, and so the variance increases.
The variability of subtracting a variable is
actually the same as adding it.
Another way of thinking of it is that
calculating the variance squares the
underlying values. Var(X + bY) is equal to
Var(X) + b
2
Var(Y), and if b is -1, this gives us
Var(X - Y). As (-1)
2
= 1, this means that
Var(X - Y) = Var(X) + Var(Y).
Q:
Can we do this if X and Y aren’t
independent?
A: No, these rules only apply if X and
Y are independent. If you need to find the
variance of X + Y where there’s dependence,
you’ll have to calculate the probability
distribution from scratch.
Q:
It looks like the same rules apply
for X + Y as X
1
+ X
2
. Is this correct?
A: Yes, that’s right, as long as X, Y, X
1
and X
2
are all independent.
Independent observations of X are different instances
of X. Each observation has the same probability
distribution, but the outcomes can be different.
If X
1
, X
2
, ..., X
n
are independent observations of X then:
E(X
1
+ X
2
+ ... + X
n
) = nE(X)
Var(X
1
+ X
2
+ ... X
n
) = nVar(X)
If X and Y are independent random variables, then:
E(X + Y) = E(X) + E(Y)
E(X - Y) = E(X) - E(Y)
Var(X + Y) = Var(X) + Var(Y)
Var(X - Y) = Var(X) + Var(Y)
The expectation and variance of linear transforms of X
and Y are given by
E(aX + bY) = aE(X) + bE(Y)
E(aX - bY) = aE(X) - bE(Y)
Var(aX + bY) = a2Var(X) + b2Var(Y)
Var(aX - bY) = a2Var(X) + b2Var(Y)