Let be a finite-dimensional inner product space over
, and
be a linear transformation (or if you’d prefer, linear operator on
). Recall that
is diagonalisable if and only if there exists a linearly independent set
of eigenvectors of
such that
.
Not all operators are diagonalisable (though all operators over will have a Jordan normal form, details here). But diagonalisable operators are convenient since we can compute
with relative ease. What’s fascinating, however, is that in the case
, if
is normal, then
is diagonalisable, i.e. there exists a basis
of
such that each vector in
is an eigenvector of
. Something nicer, in fact, happens in this case. It turns out that
must be orthonormal. These two conditions turn out to be equivalent, and proving this surprising fact is our goal for this post.
Lemma 1. Multiplication in the subset of of diagonal matrices is commutative.
Now suppose or
.
Theorem 1. Suppose is a product of linear factors in
.
is normal if and only if there exists an orthonormal (ordered) set
of eigenvectors of
such that
.
Proof. We first prove the direction . Suppose there exists an orthonormal set
of eigenvectors of
such that
. Then
is similar to the matrix
By definition of the conjugate-transpose,
Hence, since both matrices are diagonal, . We leave it as an exercise to verify that
. Hence,
implies that
, i.e.
is normal.
For the direction , we will prove using induction on
. The case
is left as an obvious exercise. Assume that
is a normal operator and suppose
. By hypothesis, the characteristic polynomial
has at least one complex root
, which means that
has at least one unit eigenvector
such that
.
Define the subspace . Since
is finite-dimensional, we can make the direct sum decomposition
. By considering dimensions,
so that . We claim that
is a normal operator. Firstly, we need to check that
, i.e. that
is
-invariant, so that
is indeed a linear operator. Secondly, we need to prove that
, so that
is indeed normal.
For the first claim, fix , so that
. We need to check that
. Since
is normal,
so that we do have , as required.
For the second claim, we first observe that for any ,
Hence, . By a similar argument, as before, for any
,
so that is
-invariant. Hence,
Therefore, is normal since
is normal. Similarly,
is a product of linear factors in
. By the induction hypothesis, there exists an orthonormal basis
of eigenvectors of
for
. Then
forms an orthonormal basis for
, as required.
Corollary 1. is normal if and only if there exists an orthonormal basis
of
consisting of eigenvectors of
such that
is diagonal.
Corollary 2. The square matrix is normal if and only if there exists a unitary matrix
such that
is diagonal.
Theorem 2. Suppose .
is self-adjoint if and only if there exists an orthonormal (ordered) set
of eigenvectors of
such that
.
Proof. For the direction , the proof in Theorem 1 yields that
so that , as required.
For the direction , suppose
is self-adjoint. Since
is normal, by Theorem 1, it suffices to check that
is a product of linear factors in
. Let
be any root of
. Then there exists
such that
Since , we must have
, yielding
. Hence,
can be written as a product of linear factors, as required.
Corollary 3. The square matrix is symmetric if and only if there exists an orthogonal matrix
such that
is diagonal.
These climactic results are known as the spectral theorems, and we will revisit them in the following crucial result: the singular value decomposition. These results cap off a theoretical study in undergraduate linear algebra.
In subsequent posts, we will explore various applications of what we have learned in other areas of mathematics.
—Joel Kindiak, 18 Mar 25, 1812H
Leave a comment