
you are here 4    233
using discrete probability distributions
Q:
 So if X and Y are games, does  
aX + bY mean a games of X and b games 
of Y?
A: aX + bY actually refers to two linear 
transforms added together. In other words, 
the underlying values of X and Y are 
changed. This is different from independent 
observations, where each game would be an 
independent observation.
Q:
 I can’t see when I’d ever want to 
use X – Y. Does it have a purpose?
A: X – Y is really useful if you want to find 
the difference between two variables.  
E(X – Y) is a bit like saying “What do you 
expect the difference between X and Y to 
be”, and Var(X – Y) tells you the variance.
Q:
 Why do you add the variances for  
X – Y? Surely you’d subtract them?
A: At first it sounds counterintuitive, 
but when you subtract one variable from 
another, you actually increase the amount 
of variability, and so the variance increases. 
The variability of subtracting a variable is 
actually the same as adding it. 
 
Another way of thinking of it is that 
calculating the variance squares the 
underlying values. Var(X + bY) is equal to 
Var(X) + b
2
Var(Y), and if b is -1, this gives us 
Var(X - Y). As (-1)
2
 = 1, this means that  
Var(X - Y) = Var(X) + Var(Y). 
 
Q:
 Can we do this if X and Y aren’t 
independent?
A: No, these rules only apply if X and 
Y are independent. If you need to find the 
variance of X + Y where there’s dependence, 
you’ll have to calculate the probability 
distribution from scratch.
Q:
 It looks like the same rules apply 
for X + Y as X
1
 + X
2
. Is this correct?
A: Yes, that’s right, as long as X, Y, X
1
 
and X
2
 are all independent.
Independent observations of X are different instances 
of X. Each observation has the same probability 
distribution, but the outcomes can be different.
If X
1
, X
2
, ..., X
n
 are independent observations of X then: 
 
  E(X
1
 + X
2
 + ... + X
n
) = nE(X) 
  Var(X
1
 + X
2
 + ... X
n
) = nVar(X) 
If X and Y are independent random variables, then: 
 
  E(X + Y) = E(X) + E(Y) 
  E(X - Y) = E(X) - E(Y) 
  Var(X + Y) = Var(X) + Var(Y) 
  Var(X - Y) = Var(X) + Var(Y)
The expectation and variance of linear transforms of X 
and Y are given by 
 
  E(aX + bY) = aE(X) + bE(Y) 
  E(aX - bY) = aE(X) - bE(Y) 
  Var(aX + bY) = a2Var(X) + b2Var(Y) 
  Var(aX - bY) = a2Var(X) + b2Var(Y)