We are now ready to ascend the mountain of proving the central limit theorem, which we will take inspiration from this paper by Calvin Wooyoung Chin.
Lemma 1. Let be a real-valued random variable with finite expectation. For any
,
This is called Markov’s inequality. Furthermore, if has finite variance, then
This is called Chebychev’s inequality.
Proof. For the first identity,
Dividing by yields the desired result. For Chebychev’s inequality, apply Markov’s inequality to the random variable
to get
Chebychev’s inequality is also responsible for our intuition about probability arising from repeated experiments.
Corollary 1. Fix and denote
. Define the i.i.d. random variables
by
, so that
. Define
Then for any ,
. In this case, we say that
in probability.
Proof. Fix . Since
and
,
By Chebychev’s inequality,
Taking ,
.
Denote and its c.d.f. by
.
Lemma 2. Let be real-valued random variables. Suppose for any thrice-differentiable
such that
are bounded,
Then for any ,
Proof. Fix and
. Since the p.d.f. of
is continuous, there exists
such that
In particular, . Define thrice-differentiable bounded functions
with bounded first, second, and third derivatives such that
By hypothesis,
Similarly, . Since
point-wise, for sufficiently large
,
so that , as required.
Theorem 1 (Central Limit Theorem). Let be i.i.d. with mean
and variance
. Then for any
,
In this case, we say that in distribution.
Proof. Fix i.i.d. that are independent of
. For each
and
, define
By direct computations, and
. The key idea is using Lemma 2 to conclude our proof. Fix any thrice-differentiable
such that
are bounded. We claim that
Fix . We aim to find
such that for
,
Applying Taylor’s theorem for the interval containing and
, there exists
between them such that
Performing algebra yields
Since is bounded, there exists
such that
By the mean value theorem, whenever ,
Consider the “good event” and its complement
By our bound on the second derivative via the mean value theorem, we can bound in the “good event”
by
On the other hand, in the “bad event” , since
for some
,
By Chebychev’s inequality,
Therefore,
Thus, selecting yields
. Plugging in the left-hand side of
,
The argument remains almost unchanged with replaced with
, so that we have the bound
Applying the triangle inequality,
Finally, applying a telescoping series,
Replace with
to complete the argument.
With that, we are done with probability…for now. There are many directions we can take from this point on. We could have proven this result using techniques involving characteristic functions or even Brownian motion, but those topics will require their entire blog sections in order to properly discuss. We conclude with the famous normal approximation to the binomial and Poisson distributions.
Corollary 1. For ,
This is the normal approximation to the binomial distribution.
Proof. For each , write
for i.i.d.
and
Then
Since the are i.i.d. with mean
and variance
, by the central limit theorem,
Corollary 2. We have
This is the normal approximation to the Poisson distribution.
Proof. Assume . Write
for i.i.d.
and
Then
Since the are i.i.d. with mean
and variance
, by the central limit theorem,
—Joel Kindiak, 31 Jul 25, 1359H
Leave a comment