MATH 2421: Lecture 23
Date: 2024-11-25 12:00:07
Reviewed:
Topic / Chapter:
summary
❓Questions
Notes
Moment Generating Function (cont.)
- 
Moment generating function
 - 
Examples
- when
 - when 
- , being independent and
 
 - 
- 👨🏫 exercise!
 
 - when 
- suppose
 - derive the 4th moment by computing 4th derivative
 - 👨🏫 tedious!
 
 - when 
- when  for 
-  as otherwise,  - somewhat impossible:
- as = frequency can't be negative!
 
 
 -  as otherwise,  - somewhat impossible:
 - when 
- , where
 - compute first
 
 - suppose , find 
- given: Poisson distribution
 
 - sum of independent Binomial r.v. w/ the same success probability: binomial
- , , being independent
 - thus
 
 - similar for Poisson r.v.
 - sum of independent normal r.v. is normal
 - find all moments of exponential dist. of parameter 
- thus:
 
 - chi-squared r.v. w/  degrees of freedom: 
- given as: 
- : independent std. normal r.v.
 
 - then :
 
 - given as: 
 
 - when  for 
 
 
Joint Moment Generating Functions
- 
Joint MGFs
- for multiple variables:
 - thus: for joint distribution of 
- exists joint cdf, pmf/pdf, and joint mgf
- joint mgf:
 
 
 - exists joint cdf, pmf/pdf, and joint mgf
 - individual mgf: can be obtained from joint mgf (like marginal)
 - 👨🏫 not required for the final exam
 
 - for multiple variables:
 - 
Multiplication rule
- 
- iff each independent
 
 - proof: too advanced for this course
- but some generalization of the original multiplication rule
 
 
 - 
 - 
Example
- let  be independent normal r.v. w/ 
- as we know that  are independent, we get:
- if we considered the original preceding as a sum of normal r.v. with 
- and indep. normal r.v. w/
 - joint MGF: uniquely determines the joint distribution
 - thus: and are independent normal r.v.
 
 
 - if we considered the original preceding as a sum of normal r.v. with 
 
 - as we know that  are independent, we get:
 
 - let  be independent normal r.v. w/ 
 
Limit Theorem
- 
Introduction
- 2 important concepts: laws of large numbers & central limit theorem
 - LLN: concerned w/ stating conditions under the average of a seq. of r.v. converges to expected average
 - CLT: concerned w/ determining conditions: where sum of large number of r.v. has probability distribution approximately normal
 
 - 
Chebyshev's inequality and weak law of large numbers
- inequalities below: for estimating tail probabilities w/ minimal information on r.v.
- e.g. its mean / variance
 
 - Markov's inequality: for non-negative r.v. and for :
 - proof: (~= first moment)
 - Chebyshev's inequality: for r.v. w/ finite mean , variance , then for :
 - proof: (~= first moment)
 - Markov's & Chebyshev's: deriving bounds on probabilities with very few information
- mean / mean & variance
 - if the actual distribution is not known: it's a decent measure!
 
 
 - inequalities below: for estimating tail probabilities w/ minimal information on r.v.
 - 
Example
- no. of items produced in factory: a r.v. w/ mean 50
- what is: probability that production exceed 75?
 - if variance of production is 25, what is the probability of production being within ?
 
 - what is: probability that production exceed 75?
 
 - no. of items produced in factory: a r.v. w/ mean 50