In Section 2.1, we saw conditions that enable us to represent a linear transformation as an upper triangular matrix. This is our first theorem about a standard form, and it puts us in a position to prove an important result that we used earlier called the Spectral Theorem.
The version of the Spectral Theorem that we saw concerns real symmetric matrices, which are square matrices for which \(A=A^T\text{.}\) This necessarily means that we are working in an inner product space so we will first extend our results on upper triangular matrices to operators on inner product spaces.
Subsection2.2.1The Schur decomposition
We will first consider complex vector spaces. In particular, suppose that \(V\) is a finite-dimensional complex inner product space and that \(T:V\to V\) is an operator on \(V\text{.}\) By the Fundamental Theorem of Algebra, we know that the minimal polynomial of \(T\) can be written as a product of linear factors:
which tells us that there is a basis \(\bcal\) in which \(\coords{T}{\bcal}\) is upper triangular. We will denote the vectors in \(\bcal\) as \(\bcal=\{\basis{\vvec}{n}\}\text{.}\)
Since \(V\) is an inner product space, we can apply the Gram-Schmidt algorithm to \(\bcal\) to form a new orthogonal basis \(\ccal\text{.}\) The vectors in \(\ccal\) will be denoted by \(\ccal=\{\basis{\wvec}{n}\}\) so that
We obtain an orthonormal basis by setting \(\uvec_j=\frac{\wvec_j}{\len{\wvec_j}}\text{.}\) Since this change of coordinates matrix is diagonal, we obtain the following result.
Theorem2.2.1.Schur decomposition.
If \(T:V\to V\) is an operator on a finite dimensional complex inner product space, then there is an orthonormal basis in which the matrix representing \(T\) is upper triangular.
This result is sometimes expressed in terms of matrices. We earlier considered orthogonal matrices, which are real matrices whose columns form an orthonormal basis. If the matrix is complex, such a matrix is called unitary.
Definition2.2.2.
A complex \(n\times n\) matrix \(U\) whose columns form an orthonormal basis for \(\complex^n\) is called unitary. Such a matrix satisfies \(U^*U=UU^*=I\text{.}\)
We can now restate the Schur decomposition in terms of unitary matrices.
Theorem2.2.3.Schur decomposition.
If \(A\) is a square complex matrix, then there is a unitary matrix \(U\) and an upper triangular matrix \(T\) such that
\begin{equation*}
A=UTU^*\text{.}
\end{equation*}
Subsection2.2.2Self-adjoint operators
When \(V\) and \(W\) are inner product spaces, a linear transformation \(T:V\to W\) has an adjoint \(T^*:W\to V\) as introduced in Subsection 1.3.3. When expressed in terms of orthonormal bases for \(V\) and \(W\text{,}\) the matrix associated to \(T^*\) is the conjugate transpose of the matrix associated to \(T\text{.}\) Or when the vector spaces are real, the matrices are simply the transpose of one another.
We will now consider operators \(T:V\to V\) on an inner product space \(V\) that are self-adjoint.
Definition2.2.4.
We say that operator \(T\) on an inner product space \(V\) is self-adjoint if \(T=T^*\text{.}\)
Theorem2.2.5.The complex spectral theorem.
If \(T:V\to V\) is a self-adjoint operator on a complex vector space \(V\text{,}\) then there is an orthonormal basis \(\bcal\) for which \(\coords{T}{\bcal}\) is diagonal. Furthermore, the eigenvalues of \(T\) are real.
Proof.
By the Schur decomposition 2.2.1, we know that there is an orthonormal basis \(\bcal\) for which \(\coords{T}{\bcal} = A\) is upper triangular. However, since \(T=T^*\text{,}\) we also know that \(A=\conj{A}^T\text{,}\) which says that \(A\) is diagonal with real entries on the diagonal.
For real inner product spaces, self-adjoint operators are represented by symmetric matrices.
Lemma2.2.6.
If \(T\) is a self-adjoint operator on an inner product space \(V\) and \(b\) and \(c\) are real numbers for which \(b^2 \lt
4c\text{,}\) then the operator
\begin{equation*}
T^2 + bT + cI
\end{equation*}
is an isomorphism.
Proof.
By the Fundamental Theorem of Linear Maps 1.2.13, we only need to show that \(\nul(T)=\{\zerovec\}\text{.}\) Therefore, we suppose that \(\vvec\) is a nonzero vector and consider
where each \(\lambda_j\in\complex\text{.}\) If \(\field=\real\) and \(T\) is a self-adjoint operator on \(V\text{,}\) we can reach a similar conclusion.
Proposition2.2.7.
If \(V\) is a real inner product space and \(T\) is a self-adjoint operator on \(T\text{,}\) then the minimal polynomial of \(T\) has the form
If \(n\gt 0\text{,}\) then \(T^2+b_1T+c_1I\) is invertible by Lemma 2.2.6. If we multiply \(p(T)\) by its inverse, we obtain another polynomial \(q\) of smaller degree for which \(q(T)=0\text{.}\)
Since the minimal polynomial \(p\) is the polynomial having the smallest degree among all polynomials for which \(p(T)=0\text{,}\) we conclude that \(m=0\) and therefore
If \(T\) is a self-adjoint operator on a real vector space \(V\text{,}\) then there is an orthonormal basis such that the matrix associated to \(T\) is diagonal.
Proof.
By Theorem 2.1.3 and Proposition 2.2.7, we know that there is a basis \(\bcal\) of \(V\) for which the matrix associated to \(T\) is upper triangular. As before, we apply the Gram-Schmidt algorithm to obtain an orthonormal basis \(\ccal\) and note that the change of coordinates matrix is upper triangular. Therefore,