We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 4
Proof the Central Lomt Theorem and Related
Back ground Trtacmetion on Moment Generating Functions
Reference’ /
Rice, TA Mathematical Sfetsties and Date Analysis
Uh
Dux bary Press, Belment, CA (12952,
The Moment-Generating Function
‘This section develops and applies some of the properties of the moment-generating
function. It tums out, despite its unlikely appearance, to be a very ‘useful tool that
can dramatically simplify certain calculations.
The moment generating function (mgf) of a random variable X is M()=
E(¢*) if the expectation is defined. In the discrete case,
MO = PO
and in the continuous case,
Mw) = f f(x) de
“The expectation, and hence the moment-generating function, may of may not exs
for any particular value of In the continuous case the existence of the expects4.5 The Moment-Generating Function \X}
depends on how rapidly the tails of the density decrease; for example, since the tails,
of the Cauchy density die down at the rate x~?, the expectation does not exist for
any ¢ and the moment-generating function is undefined. Since the tails of the normal
density die down at the rate e~", the integral converges for al .
PROPERTY A
If the moment-generating function exists for ¢ in an open interval containing zero, it
uniquely determines the probability distribution. =
‘We cannot prove this important property here—its proof depends on properties of
the Laplace transform. Note that Property A says that if two random variables have
the same mgf in an open interval containing zero, they have the same distribution
For some problems, we can find the mgf and then deduce the unique probability
distribution corresponding to it.
‘The rth moment of a random variable is E(X’) if the expectation exists. We have
already encountered the first and second moments earlier in this chapter, that is,
E(X) and E(X*), Central moments rather than ordinary moments are often used: The
rth central moment is E{{X — E(0))’}. The variance is the second central moment
and is a measure of dispersion about the mean. The third central moment, called
the skewness, is used as a measure of the asymmetry of a density or a frequency
function about its mean; if a density is symmetric about its mean, the skewness is
zero (see Problem 72 at the end of this chapter). As its name implies, the moment-
generating function has something to do with moments. To see this, consider the
continuous case:
MO = f $0) de
‘The derivative of M(t) is
mM
ap
> e* f(x) de
$f “ee
In ean be show that eitferentation and iteration canbe interchanged, so that
M()= f xe"f(x) de
and
M0) f af (x) dx = E(X)
Differentiating r times, we find
MO) =E)
It can further be argued that if the moment-generating function exists in an interval
containing zero, then so do all the moments. We thus have the following property.Proof
5.3 Comergence in Distribution andthe Cental Limit Theorem 18
‘The central limit theorem is concerned not with the fact that the ratio S,/n con-
verges to yx but with how it fluctuates around jz. To analyze these fluctuations, we
standardize:
S,- mu
n= OT
It is easily verified that Z, has mean 0 and variance 1. The central limit theorem
states that the distribution Of Z,, converges to the standard normal distribution.
THEOREM B_ Central Limit Theorem
Let X,,X,... be a sequence of independent random variables having mean 0 and
variance o? and the common distribution function F and moment-generating func-
tion M defined in a neighborhood of zero. Let
520%
Jin? (Sigss) 00, -w 0. Since E(X) = 0, M'(0) = 0, and M"(0) = 07. As n>
00, t/(0 Vn) —> 0, and
(si) =144 (sim) +e
ayn} 27 on,
where ¢,/(F /(no”)) + as n + 00. We thus have
;
Itcan be shown that if @, —> a, then
and
«
li 1
im, (14
From this result, it follows that
Mz (0 2 asn—> 0EXAMPLE C
EXAMPLE D
M0 Chapter 5: Limit Theorems
where exp(/?/2) is the mgf of the standard normal distribution, as was to be
shown, #
‘Theorem B is one of the simplest versions of the central limit theorem; there
are many central limit theorems of various degrees of abstraction and generality. We
have proved Theorem B under the assumption that the moment-generating functions
exist, which is a rather strong assumption. By using characteristic functions instead,
wwe could modify the proof so that it would only be necessary that first and sec~
ond moments exist. Further generalizations weaken the assumption that the X; have
the same distribution and apply to linear combinations of independent random vari-
ables. Central limit theorems have also been proved that weaken the independence
assumption and allow the X; to be dependent but not “too” dependent. Central limit
theorems are still an active area of research in probability theory.
For practical purposes, especially for statistics, the limiting result in itself is not
of primary interest. Statisticians are more interested in its use as an approximation
with finite values of n. It is impossible to give a concise and definitive statement
‘of how good the approximation is, but some general guidelines are available, and
‘examining special cases can give insight. How fast the approximation becomes good
depends on the distribution of the summands, the X;. If the distribution is fairly
symmetric and has tails that die off rapidly, the approximation becomes good for
relatively small values of n, If the distribution is very skewed or if the tails die down
very slowly, a larger value of n is needed for a good approximation, The following
examples deal with two special cases.
‘Since the uniform distribution on (0, 1] has mean 4 and variance qb, the sum of
12 uniform random variables, minus 6, has mean 0 and variance 1. The distribution
of this sum is quite close to normal; in fact, before better algorithms were devel-
oped, it was commonly used in computers for generating normal random variables
from uniform ones. It is possible to compare the real and approximate distributions
analytically, but we will content ourselves with a simple demonstration. Figure 5.1
shows a histogram of 1000 such sums with a superimposed normal density function.
The fit is surprisingly good, especially considering that 12 is not usually regarded as
alarge value of. #
‘The sum of nm independent exponential random variables with parameter A = 1 fol-
lows a gamma distribution with = 1 and a =n (Example F in Section 4.5). The
‘exponential density is quite skewed; therefore, « good approximation of a standard-
ized gamma by a standardized normal would not be expected for small n. Figure
5.2 shows the caf’s of the standard normal and standardized gamma distributions for
increasing values of n, Note how the approximation improves as n increases. *
ST - Joseph'S College of Engineering, Chennai-119 ST - Joseph'S Institute of Technology, Chennai-119 Ma6453 - Probability and Queueing Theory Unit I Random Variables Formulae Sheet