Section 5.4 The spectral theorem
Among the many takeaways from Section 4.5 is the simple fact that not all matrices are diagonalizable. In principle Theorem 4.5.13 gives a complete answer to the question of diagonalizability in terms of eigenspaces. However, you should not be mislead by the artificially simple examples treated in Section 4.5. In practice even the determination (or approximation) of the distinct eigenvalues of an matrix poses a very challenging computational problem as gets large. As such the general question of whether a matrix is diagonalizable remains an intractable one. This makes all the more welcoming the main result of this section: all symmetric matrices are diagonalizable! This surprising fact is a consequence of the spectral theorem for self-adjoint operators: a result which itself fits into a larger suite of spectral theorems that treat the diagonalizability of various families of linear transformations of inner product spaces (both finite and infinite dimensional).
Subsection 5.4.1 Self-adjoint operators
Though we are mainly interested in the diagonalizability of symmetric matrices, our arguments are made more elegant by abstracting somewhat to the realm of linear transformations of inner product spaces. In this setting the appropriate analogue of a symmetric matrix is a self-adjoint linear transformation.
The next theorem makes explicit the connection between self-adjoint operators and symmetric matrices.
Theorem 5.4.2. Self-adjoint operators and symmetry.
Let be a finite-dimensional inner product space, let be a linear transformation, and let be an orthonormal ordered basis of The following statements are equivalent.
is self-adjoint. is symmetric.
Proof.
Let We have
It follows that
The last equality in this chain of equivalences states that satisfies property (5.4.1) for all elements of Not surprisingly, this is equivalent to satisfying the property for all elements in (See Exercise 5.4.3.10.) We conclude that is symmetric if and only if is self-adjoint.
Corollary 5.4.3. Self-adjoint operators and symmetry.
Let The following statements are equivalent.
is symmetric. is self-adjoint with respect to the dot product.
Proof.
Since where is the standard ordered basis of and since is orthonormal with respect to the dot product, it follows from Theorem 5.4.2 that statements (1) and (2) are equivalent. Statements (2) and (3) are equivalent since by definition for all
The next result, impressive in its own right, is also key to the induction argument we will use to prove Theorem 5.4.8. A proper proof would require a careful treatment of complex vector spaces: a topic which lies just outside the scope of this text. The “proof sketch” we provide can easily be upgraded to a complete argument simply by justifying a few statements about and its standard inner product.
Theorem 5.4.4. Eigenvalues of self-adjoint operators.
Proof sketch of Theorem 5.4.2.
Pick an orthonormal ordered basis of and let By Theorem 5.4.2, is symmetric. To prove that all roots of the characteristic polynomial are real, we make a slight detour into complex vector spaces. The set
of all complex -tuples, together with the operations
and
where forms what is called a vector space over . This means that satisfies the strengthened axioms of Definition 3.1.1 obtained by replacing every mention of a scalar with a scalar Additionally, the vector space has the structure of a complex inner product defined as
where denotes the complex conjugate of for each Essentially all of our theory of real vector spaces can be “transported” to complex vector spaces, including definitions and results about eigenvectors and inner products. The rest of this argument makes use of this principle by citing without proof some of these properties, and this is why it has been downgraded to a “proof sketch”.
We now return to and its characteristic polynomial Recall that we want to show that all roots of are real. Let be a root of The complex theory of eigenvectors implies that there is a nonzero vector satisfying On the one hand, we have
using properties of our complex inner product. On the other hand, since it is easy to see that Corollary 5.4.3 extends to our complex inner product: i.e.,
for all Thus
(In the last equality we use the fact that our complex inner product satisfies for any and vectors ) It follows that
Since we have (another property of our complex inner product), and thus Since a complex number satisfies if and only if if and only if is real, we conclude that is a real number, as claimed.
Corollary 5.4.5. Eigenvalues of self-adjoint operators.
If is a self-adjoint operator on a finite-dimensional inner product space then has an eigenvalue: i.e., there is a and nonzero such that
Proof.
The corollary follows from Theorem 5.4.4 and the fact that the eigenvalues of are the real roots of its characteristic polynomial (4.4.25).
From Theorem 5.4.4 and Corollary 5.4.3 it follows that the characterisitic polynomial of any symmetric matrix must factor as a product of linear terms over as illustrated by the next two examples.
Example 5.4.6. Symmetric matrices.
Solution.
Given a symmetric matrix
we have
Using the quadratic formula and some algebra, we see that the roots of are given by where (using the quadratic formula)
Since we see that both these roots are real. Thus where
Example 5.4.7. Symmetric matrix.
Verify that the characteristic polynomial of the symmetric matrix
factors into linear terms over
Solution.
The characteristic polynomial of is We can use the quadratic equation to solve for yielding
We conclude that or and thus or It follows that
Subsection 5.4.2 The spectral theorem for self-adjoint operators
Our version of the spectral theorem concerns self-adjoint linear transformations on a finite-dimensional inner product space. It tells us two remarkable things: (a) every such linear transformation has an eigenbasis (and hence is diagonalizable); and furthermore, (b) the eigenbasis can be chosen to be orthogonal, or even orthonormal.
Theorem 5.4.8. Spectral theorem for self-adjoint operators.
Let be a finite-dimensional vector space, and let be a linear transformation. The following statements are equivalent.
is self-adjoint. is diagonalizable and eigenvectors with distinct eigenvalues are orthogonal. has an orthogonal eigenbasis. has an orthonormal eigenbasis.
Proof.
We will prove the cycle of implications
Implication: .
Assume is self adjoint. First we show that eigenvectors with distinct eigenvalues are orthogonal. To this end, suppose we have and where Using the definition of self-adjoint, we have
We now prove by induction on that if is self-adjoint, then is diagonalizable. The base case is trivial. Assume the result is true of any -dimensional inner product space, and suppose By Corollary 5.4.5 there is a nonzero with Let Since we have The following two facts are crucial for the rest of the argument and are left as an exercise (5.4.3.11).
- For all
we have and thus by restricting to we get a linear transformation - The restriction
is self-adjoint, considered as a linear transformation of the inner product space Here the inner product on the subspace is inherited from by restriction.
Now since and is self-adjoint, we may assume by induction that has an eigenbasis We claim that is an eigenbasis of Since by definition for all we see that the vectors are also eigenvectors of and thus that consists of eigenvectors. To show is a basis it is enought to prove linear independence, since Suppose we have
for scalars Taking the inner product with we have :
It follows that we have
and thus for all since is linearly independent. Having proved that is an eigenbasis, we conclude that is diagonalizable.
Implication: .
Let be the distinct eigenvalues of Since is assumed to be diagonalizable, following Procedure 4.5.14 we can create an eigenbasis by picking bases of each eigenspace and combining them. We can always choose these bases so that each is orthogonal. When we do so, the assembled will be orthogonal as a whole. Indeed given any two elements of if both vectors are elements of for some then they are orthogonal by design; furthermore, if is an element of basis and is an element of basis with then they are eigenvectors with distinct eigenvalues, and hence orthogonal by assumption!
Implication: .
This is easy to see since an orthonormal eigenbasis can be obtained from an orthogonal eigenbasis by scaling each element by the reciprocal of its norm.
Implication: .
Assume is an orthonormal eigenbasis of Since is an eigenbasis, is a diagonal matrix, and hence symmetric. Since is orthonormal with respect to the dot product, we conclude from Theorem 5.4.2 that is self-adjoint.
An operator that admits an orthogonal (and hence an orthonormal) eigenbasis is called orthogonally diagonalizable.
Definition 5.4.9. Orthogonally diagonalizable.
Let be a finite-dimensional inner product space. A linear transformation is orthogonally diagonalizable if there exists an orthogonal (equivalently, an orthonormal) eigenbasis of
This new language affords us a more succinct articulation of Theorem 5.4.8: to be self-adjoint is to be orthogonally diagonalizable. Think of this as a sort of “diagonalizable+” condition.
Mantra 5.4.10. Self-adjoint mantra.
As an immediate consequence of Theorem 5.4.8, we have the following result about symmetric matrices.
Corollary 5.4.11. Spectral theorem for symmetric matrices.
is symmetric. is diagonalizable and eigenvectors with distinct eigenvalues are orthogonal with respect to the dot product. is orthogonally diagonalizable.
Proof.
By Corollary 5.4.3 we have symmetric if and only if is self-adjoint with respect to the dot product. Statements (1)-(3) are seen to be equivalent by applying Theorem 5.4.8 to (with respect to the dot product). Let be the standard basis of We see that (4) is equivalent to (3) by observing that is an orthonormal eigenbasis of if and only if the matrix obtained by placing the elements of as columns is orthogonal and diagonalizes
The process of finding matrices and satisfying (5.4.3) is called orthogonal diagonalization. A close look at the proof of Theorem 5.4.8 gives rise to the following orthogonal diagonalization method for matrices.
Procedure 5.4.12. Orthogonal diagonalization.
- Let
be the ordered basis obtained by concatenating the orthonormal bases computed in (1). This is an orthonormal basis of eigenvectors. It follows that the matrixis orthogonal (i.e., ), and the matrix is diagonal.
Example 5.4.13. Orthogonal diagonalization.
The symmetric matrix
Solution.
First we factor Looking at the constant term we see that the only possible integer roots of are It is easily verified that and polynomial division yields the factorization Further factorization of gives us
Next we compute orthonormal bases of the eigenspaces and yielding
Assembling these bases elements into the orthogonal matrix
we conclude that where
Observe that the two eigenspaces and of the matrix in Example 5.4.13 are orthogonal to one another, as predicted by the spectral theorem. Indeed, is the line passing through the origin with direction vector and is its orthogonal complement, the plane passing through the origin with normal vector Figure 5.4.14 depicts the orthogonal configuration of the eigenspaces of this example. This is an excellent illustration of what makes the diagonalizability of symmetric matrices (and self-adjoint operators) special. Keep it in mind!
Do not overlook the reverse implication of equivalence (5.4.2). As the next example illustrates, we can show an operator is self-adjoint by examining the geometry of its eigenspaces.
Example 5.4.15. Orthogonal projections are self-adjoint.
Let be a finite-dimensional inner product space, let be a subpsace of and let be orthogonal projection onto Prove that is self-adjoint.
Solution.
By Theorem 5.4.8 it suffices to show that is orthogonally diagonalizable. According to Exercise 5.3.6.20 we have
Equivalently, and are the 1- and 0-eigenspaces of respectively. Since we conclude that is diagonalizable. Since clearly and are orthogonal, we conclude that is in fact othogonally diagonalizable, hence self-adjoint.
Exercises 5.4.3 Exercises
WeBWork Exercises
1.
Exercise Group.
Orthogonally diagonalize the given symmetric matrix following Procedure 5.4.12: i.e. find a diagonal matrix and orthogonal matrix satisfying
10.
11.
Let be a finite-dimensional inner product space, let be a self-adjoint operator, and let be a subspace of
- By (a), restricting
to defines a linear transformationProve that is self-adjoint. Here the inner product on the subspace is inherited from by restriction.
12.
Assume is symmetric and orthogonal. Prove that the characteristic polynomial of factors as for some nonnegative integers In particular, the eigenvalues of are among and
Exercise Group.
13.
14.
15.
16.
17.
- First graph
and then graph using the result of (a). What type of conics (parabolas, ellipses, hyperbolas) are and ?