We are now justified in adding random variables. For instance, if , what is the distribution of
? Unfortunately, in the most extreme cases, this answer is trivial.
Lemma 1. Let and
. Then
and
.
Proof. We observe that for any , since
, by a change of variables,
Therefore, for any ,
Therefore, so that
.
Here, it is clear that is entirely dependent—
correlated even—to the random variable
. We got a rather useless answer to our question in this case. If instead we swung to the other extreme, namely
being independent, we don’t have a general formula for any
. Nevertheless, the independent case proves to be far more useful in reality—for instance, the exam scores of two students are effectively independent barring (sufficiently drastic) externalities.
Therefore, let’s discuss what it means for two random variables to be independent. We have previously seen that two events
are called independent if
.
Lemma 2. The set forms a
-algebra, called the
-algebra generated by
.
Given two -algebras
, also known as sub-
-algebras of
, we say that
are independent if for any
,
are independent.
Definition 1. Two random variables are said to be independent if
are independent.
Suppose and
, where
denotes either the counting measure
or the Lebesgue measure
. Let
be the product of two copies of
. Suppose
.
Theorem 1. are independent and
if and only if
where the integrand in the right-hand side is interpreted to mean
Proof. Consider the cumulative distribution functions
By the Radon-Nikodým theorem,
In the direction , by the Fubini-Tonelli theorem,
establishing independence.
In the direction , we similarly use the Fubini-Tonelli theorem to obtain
so that .
Henceforth, when are independent and
and
, we assume
so that
is a meaningful quantity.
Corollary 1. Define
Then are independent and
if and only if for any
,
Finally, let’s add these random variables together.
Theorem 1. If are independent
-valued random variables with density functions
, then
is a random variable with density function
where the right-hand side denotes convolution with respect to :
If is the counting measure, we get
If is the Lebesgue measure, we get
Proof. Denote so that for any fixed
,
if and only if
. For any
,
Finally, let’s concretely add two independently normally distributed random variables.
Theorem 2. If and
are independent, then
.
Proof. Write and
. Then
where are independent. It suffices to prove that
By construction, for any , if
, then
Defining ,
By algebruh,
for carefully calculated constants , in particular, with
Denoting , the integral then simplifies to
Therefore, denoting so that
,
Therefore, , as required.
In practice, to compute probabilities involving normal distributions, we define
and tabulate commonly used approximate values of for
, known in the statistics community as a
-table. By symmetry,
. For any
, since
, we can reduce the computation to
By induction, we have the following sampling distribution .
Corollary 2. For independent, identically distributed random variables ,
The central limit theorem claims that even if are not normally distributed,
converges in distribution to
. For the limiting distribution to be constant, the equivalent claim is that
converges in distribution to
.
We have previously stated the special case and aim to prove it properly using techniques in stochastic calculus. If we have this result, we are emboldened to carry out hypothesis tests, which are commonplace in the STEM fields as well as the data-driven social sciences. We will digress to this application before we dive right back into our ascent toward the central limit theorem.
—Joel Kindiak, 23 Jul 25, 2257H
Leave a comment