29 SPECTRAL THEOREM
29.1. Eigenvalues and eigenvectors of symmetric matrices
Let be a square, symmetric matrix. A real scalar is said to be an eigenvalue of if there exists a non-zero vector such that
The vector is then referred to as an eigenvector associated with the eigenvalue . The eigenvector is said to be normalized if . In this case, we have
The interpretation of is that it defines a direction along behaves just like scalar multiplication. The amount of scaling is given by . (In German, the root ‘‘eigen’’, means ‘‘self’’ or ‘‘proper’’). The eigenvalues of the matrix are characterized by the characteristic equation
where the notation refers to the determinant of its matrix argument. The function, defined by , is a polynomial of degree called the characteristic polynomial.
From the fundamental theorem of algebra, any polynomial of degree has (possibly not distinct) complex roots. For symmetric matrices, the eigenvalues are real, since when , and is normalized.
29.2. Spectral theorem
An important result of linear algebra called the spectral theorem, or symmetric eigenvalue decomposition (SED) theorem, states that for any symmetric matrix, there are exactly (possibly not distinct) eigenvalues, and they are all real; further, that the associated eigenvectors can be chosen so as to form an orthonormal basis. The result offers a simple way to decompose the symmetric matrix as a product of simple transformations.
Theorem: Symmetric eigenvalue decomposition
We can decompose any symmetric matrix with the symmetric eigenvalue decomposition (SED)
where the matrix of is orthogonal (that is, ), and contains the eigenvectors of , while the diagonal matrix contains the eigenvalues of . |
Here is a proof. The SED provides a decomposition of the matrix in simple terms, namely dyads.
We check that in the SED above, the scalars are the eigenvalues, and ‘s are associated eigenvectors, since
The eigenvalue decomposition of a symmetric matrix can be efficiently computed with standard software, in time that grows proportionately to its dimension as .
Example: Eigenvalue decomposition of a symmetric matrix.
29.3. Rayleigh quotients
Given a symmetric matrix , we can express the smallest and largest eigenvalues of , denoted and respectively, in the so-called variational form
For proof, see here.
The term ‘‘variational’’ refers to the fact that the eigenvalues are given as optimal values of optimization problems, which were referred to in the past as variational problems. Variational representations exist for all the eigenvalues but are more complicated to state.
The interpretation of the above identities is that the largest and smallest eigenvalues are a measure of the range of the quadratic function over the unit Euclidean ball. The quantities above can be written as the minimum and maximum of the so-called Rayleigh quotient .
Historically, David Hilbert coined the term ‘‘spectrum’’ for the set of eigenvalues of a symmetric operator (roughly, a matrix of infinite dimensions). The fact that for symmetric matrices, every eigenvalue lies in the interval somewhat justifies the terminology.