The 21st century mathematician Rick Durrett claimed in his lecture notes that measure theory ends and probability begins with the definition of independence. We explore that idea in this topic, as well as its generalisation in the form of Bayes’ theorem.
Let be a probability space.
Lemma 1. Let be any event with positive probability:
. Then the map
is a well-defined probability measure. In particular,
if
, and
if
.
Equivalently,
Proof. It suffices to prove that , which holds since
implies
Corollary 1. The law of total probability then reduces to the following: if and each
has positive probability, then
What the quantity as defined in Lemma 1 measures, roughly speaking, is the probability that the event
occurs under the assumption that the event
occurs. This model fits the conclusion of Corollary 1 quite well: to evaluate
, we condition on each of the varied
, then evaluate
for each
, followed by summing up via the addition principle. Furthermore, we recover the following intuitions:
- if
, then any outcome
automatically yields
,
- if
, then any outcome in
automatically yields
.
In a sense, these two scenarios are the extremes: the first means that is totally included in
, and the second means that
is totally excluded from
. But could there be a middle ground? What if, in a sense, event
“does not care” about event
? Assuming that
, this observation amounts to the equality
The equation on the left means that regardless of whether the event occurs or not, the event
will hold with the same probability. The equation on the right is equivalent to the equation on the left if
, but deserves its own attention since it handles the situation where
. We use this definition for the notion of independence:
Definition 1. Two events are independent if
Example 1. Flip a fair coin twice so that its sample space is
Let denote the discrete
-algebra and
denote the uniform counting measure
on
.
Let denote the event that the first coin is a Head, and
denote the event that the second coin is a Head. Then
Thus, the events and
are independent, and misunderstanding such may lead to financial ruin.
Independence of events is an insanely useful idea in probability theory, especially in the sense of independent random variables, which we will explore in future write-ups. In fact, many kinds of events are independent:
Theorem 1. We have the following independence (or not) properties:
- for any
,
and
are independent,
- if
have positive probability, then they cannot be mutually exclusive and independent at the same time,
- if
has positive probability, then
cannot be independent.
Furthermore, if are independent, the following pairs of events are independent as well:
Proof. We illustrate the proof of being independent and leave the rest as exercises. Since
are independent, we use finite additivity to obtain
Since , we have
, so that
The reality, however, is that in most situations, the quantity is nontrivial. Furthermore, the quantity seems trivial if, at least sequentially, the event
either occurs (or not), and then the event
occurs (or not).
For example, could indicate the event “I suspect that this patient has COVID-19” and
could indicate the event “the patient tests positive for COVID-19”. Since we are finite human beings, there is no harm assuming that
and
. Now the quantity
has a natural interpretation—it measures the probability that, under the assumption that a patient has COVID-19, the patient tests positive for the virus. Summarised using intuitive terms,
measures the sensitivity of the COVID-19 test.
In a similar vein (no pun intended), measures the specificity of the COVID-19 test. The quantities
and
then measure the false positive rate and false negative rates respectively, and are connected to the sensitivity and specificity measures as follows:
Now we present our case-study that generalises nicely into Bayes’ theorem.
Example 2. Suppose of the population has been infected with COVID-19, and your COVID-19 test kit has a sensitivity of
and a specificity of
. Given that a patient tests positive using your test kit, what is the probability, in terms of
, that this patient actually has COVID-19?
Solution. Using as our events, we observe that
We are interested in the quantity . Nothing in our problem-solving arsenal helps us at the moment, except perhaps for
. We note that
, so that
Doing some algebruh and substituting relevant values,
Now taking advantage of the law of total probability,
The fascinating observation is that ,
, and
. This means if we know that
of the population has been infected with COVID-19, and assuming constant sensitivity and specificity of the COVID-19 test we tested positive for it, there is a
chance that we have been infected with the virus too. For better or for worse, these data points inform governments on various public health policies and measures taken to curb the virus and “flatten the curve“.
Our computations above prove Bayes’ theorem:
Theorem 2 (Bayes). Given events with positive probability,
Furthermore, if and each
has positive probability, then
There is more to discuss on Bayes’ theorem, but we conclude with one more application. We notice that if we are given a constant , then
Then the higher the value of and
, the higher the value of
. This connection is often used in Bayesian statistics to model an update of belief: given a prior probability
, the posterior probability is the new probability
given that we have new evidence, namely
.
We need to introduce one more key star player of probability, namely that of random variables. This we will do next time.
—Joel Kindiak, 26 Jun 25, 1847H
Leave a comment