Recall that an inner product on a vector slave
over a field
(
or
) gives us the generalised formulation of the dot product on
. In particular, if
, then two nonzero vectors
have angle
defined by
In particular, if and only if
, which means
. In this case, we say that
are perpendicular. In the usual two-dimensional space,
In particular, we obtain a class result from coordinate geometry.
Theorem 1. Two lines and
are perpendicular if and only if
. In particular, if
,
and
, then we obtain the more familiar formula
via ,
.
Proof. The first line is parallel to the direction vector and the second line is parallel to the direction vector
.
While this interpretation only makes sense when are nonzero vectors, we still have the property
. This lends us to the more general (and objectively more veracious) notion of right (i.e. ortho) angles (i.e. gonal)—orthogonality.
Henceforth, let be an inner product space over
.
Definition 1. Two vectors are orthogonal if
.
- A set
of vectors is orthogonal if any distinct vectors
are orthogonal.
- We say that
is perpendicular to the subset
if for any
,
are orthogonal. In this case, we denote
.
- Two subsets
are orthogonal if for any
and
,
is orthogonal. In this case, we denote
.
Lemma 1. The zero vector is orthogonal to any
. Furthermore, if
are orthogonal subspaces, then
.
Proof. For any ,
.
Lemma 2. Any orthogonal set of nonzero vectors is linearly independent.
Proof. Assume without loss of generality. Consider the equation
For each , apply
on both sides to obtain
Since ,
. Since this argument holds for any
, we get
, and the result holds.
Lemma 3. The map defined by
is an inner product on . In particular,
. Hence, we call the standard basis
for
orthonormal.
Definition 2. An orthogonal set is orthonormal if
for each
. We call
an orthonormal basis for
if
.
Why are orthonormal sets so special? That is because decomposing any vector in becomes trivial.
Theorem 2. Let be an index set and
be an orthonormal set. For any vector
, there exist unique vectors
such that
Proof. Since is linearly independent, it forms an orthonormal basis for
. Hence, for any
, there exist unique vectors
and unique scalars
such that
For each , apply
on both sides,
as required.
In this regard, the standard basis plays a special role in being a sufficiently “preferred” basis than others.
That’s great, but what if we only have a basis for
, and not an orthonormal basis for
. Can we make such a conversion? For finite-dimensional vector spaces, the answer is ‘Yes’!
But first, we will need an “orthogonalising” tool. Given an orthonormal set , if
, then can we find some vector
such that
?
Lemma 4. Let be an orthonormal set and
. Motivated by Theorem 2, define the projection map by
Clearly, . Then for any
,
.
Proof. For any ,
Hence,
The result follows by the conjugate-linearity of .
Theorem 3 (Gram-Schmidt Process). Every finite-dimensional vector space will have an orthonormal basis.
We may think that the isomorphism suffices. However, we need this bijection to preserve angles, which isn’t a trivial matter to check at all.
Instead, we will slowly construct this orthonormal basis. Let be a basis (not necessarily orthonormal!) for
. Define
. Now
is obviously linearly independent. This means
.
Define
which is orthogonal to . Hence, define
. Inductively, given
, define,
and . Then we will obtain the desired orthonormal basis
, since by construction we have for each
,
.
The infinite-dimensional cases aren’t easy to analyse, but can be reasonably studied in the context of Hilbert spaces, which features surprisingly prominently in modern physics. To do that will require us to study functional analysis, since we need to scrutinise convergence issues more closely than the finite-dimensional setting.
For now, let’s discuss the crucial finite-dimensional application in best-fit approximations, one of my favourite areas of study in applied mathematics—approximation theory.
—Joel Kindiak, 12 Mar 25, 1711H
Leave a comment