Professional Documents
Culture Documents
Complements
Either A occurs, or it does NOT occur: Pr( AC ) = 1 − Pr( A).
Unions
Pr( A ∪ B ) = Pr( A) + Pr( B ) − Pr( A ∩ B ).
You also must the know the extension of this formula to three random variables, which is
De Morgan’s Laws
1. Pr(( A ∪ B ) C ) = Pr( AC ∩ B C )
2. Pr( AC ∪ B C ) = Pr(( A ∩ B ) C )
Conditional probability
Pr( A ∩ B )
Pr( A | B ) = , which can also be expressed as Pr( A | B ) ⋅ Pr( B ) = Pr( A ∩ B ) .
Pr( B )
Bayes’ Theorem
Pr( B | A) Pr( A)
Pr( A | B ) =
Pr( B | A) Pr( A) + Pr( B | AC ) Pr( AC )
Expected Value
There are two ways to calculate the expected value. The basic, most direct route is one of
the things you have to know:
But there’s also another way to calculate the expected value that can be faster, based on
the information you are given in the problem.
For a discrete distribution:
∞
E( X ) = ∑ Pr( X ≥ x )
x = −∞
You have to know the basic way to calculate the expected value; the other method is also
nice to know and can save you valuable time in the heat of the exam.
Variance
The variance of a random variable X, denoted Var(X), is given by the formula
Var ( X ) = E ( X 2 ) − E ( X ) 2
Covariance
The covariance between two variables is Cov ( X , Y ) = E ( XY ) − E ( X ) E (Y ) .
Also, you need to know that Var ( X + Y ) = Var ( X ) + Var (Y ) − 2Cov ( X , Y ), and that
Var ( aX + b) = a 2Var ( X ) and Cov ( aX , bY ) = abCov( X , Y ).
The standard deviation, denoted σ X , is simply the square root of the variance,
σ X = Var ( X ) .
Coefficient of variation: This is simply the ratio of the standard deviation to the mean,
σX
that is .
E( X )
Double Expectation
E(X) = EY(E(X|Y=y))
Var(X)=VarY(E(X|Y=y))+EY(Var(X|Y=y))
Probability Distributions
Mean, median, and mode
The three most tested probability distributions are the uniform, exponential, and Poisson.
You also need to know the binomial, geometric, the negative binomial, and the
hypergeometric.
The mode is equal to lambda, rounded down to the nearest integer. For example, a
Poisson distribution with mean equal to 3.2 has a mode equal to 3. A Poisson with mean
equal to 3 also has mode equal to 3.
Also good to know is that the sum of two Poisson distributions with means λ1 and λ2 is a
Poisson distribution with mean = λ1 + λ2.
It gets a little trickier if two Poisson distributions or more are involved. A shortcut that
can save you a significant amount of time is recognizing that the sum of two or more
Poisson distributions is also a Poisson distribution. For example, supposed that you are
asked the following question:
A business models the number of customers for the first week of each month as a Poisson
distribution with mean = 3, and for the second week of each month as a Poisson
distribution with mean = 2. What is the probability of having exactly two customers in
the first two weeks of a month? The long way to do this is to figure out all the different
combinations –
The easy way to do this is to use the fact that the sum of two Poisson distributions is also
Poisson. So the sum of the Poisson distributions from weeks one and two is Poisson with
5 2 e −5
mean = 5. The probability of exactly two customers is ≈ 0.084.
2!
f ( x ) = λe − λ x
1 1
The mean equals , and the variance equals 2 . This is an important distinction from
λ λ
the Poisson, where the mean is equal to the variance. For the exponential, the mean is
equal to the standard deviation, so the variance is equal to the mean squared.
Some useful integration shortcuts that can save you valuable time on the exam:
∞ x a
1 − −
∫a θ e θ dx = e θ
∞ x a
1 − −
∫a x θ e θ dx = (a + θ )e θ
∞ x a
1 −θ −
∫x e dx = (( a + θ ) 2 + θ 2 ) e θ
2
a
θ
1 with prob. p
X =
0 with prob.1 − p = q
E(X) = p
Var(X) = p(1-p)
Binomial Distribution
n!
f ( n, k , p ) = C ( k , n ) p k (1 − p ) n −k where C ( k , n ) = .
k! ( n − k )!
Mean = np, Variance = np(1-p)
Geometric Distribution
Perform Bernoulli trials until success, then stop and count the total number of trials – this
is the geometric random variable.
The tricky part about this is that there can be two different formulations, based on
whether you count the number of trials before the first success, or the number of failures
before the first success.
k
E (n) =
p
1 1
Var ( n ) = k 2 −
p p
Hypergeometric Distribution
Used for sampling without replacement. Finite population with n objects, k are special, n-
k are not. If m objects are chosen at random, the probability that out of m, x are special is
k! ( n − k )!
x! ( k − x )! ( m − x )! ( n − k − ( m − x ))!
f ( x) =
n!
m! ( n − m )!
It looks a little complicated but once you’ve worked several problems, this is not too
hard.
Normal Distribution
Continuity correction factor for binomial or Poisson or uniform approximations:
k + 0.5 − µ k − 0.5 − µ
Pr( X = k ) = Pr( k − 0.5 < Y < k + 0.5) = Φ − Φ
σ σ
σ Y2|X = x = (1 − ρ 2 )σ Y2
Lognormal Distribution
ln y − µ ln y
F ( y ) = Pr(Y ≤ y ) = Pr(ln Y ≤ ln y ) = Φ
σ
ln y
1 2
µln y + σ ln y
E (Y ) = e 2
σ2
Var (Y ) = E (Y ) 2 e ln y − 1
Other distributions on the syllabus include the beta, the Pareto, the Chi-Square, and the
Weibull. I have not presented them here.
Marginal Density
f ( x ) = ∫ f ( x, y )dy
Y |X
Order Statistics
f X ( k ) ( x ) = nf ( x )C ( k − 1, n − 1) F ( x ) k −1 (1 − F ( x )) n −k
Conditional Density
f ( x, y )
f X |Y ( x | y ) =
f ( y)
E (Y | X = x ) = ∫y f
Y |X
Y |X ( y | x )dy
MaX(t)=MX(at)
Mb(t)=ebt
MX(0)=1
1
MGF for Exponential:
1−θ t
n
1
MGF for Gamma:
1 − θ t
Note that based on a comparison of the MGFs for the exponential and gamma
distributions, it’s easy to see that the gamma is the sum of n exponential distributions.
Joint MGFs:
M X ,Y ( s, t ) = E e sX + tY
∂ n +m
E ( X nY m ) = M X ,Y ( 0,0 )
∂ n s∂ m t
M X ,Y ( s,0) = M X ( s )
M X ,Y (0, t ) = M Y (t )
M X ,Y ( s, t ) = M X +Y (t )
Chebyshev’s Theorem
σ2
Pr ( | X − µ | ≥ c ) ≤
c
Benefit Distributions
E(X) = qE(B)
Cov ( X , Y )
Correlation coefficient: ρ ( X , Y ) =
σ XσY