Diagonalisation of symmetric and Hermitian matrices
In this last section we discuss properties of symmetric and Hermitian matrices which have applications in classical physics and quantum mechanics. The theory outlined for Hermitian matrices is closely related to Sturm-Liouville theory which defines properties of solution sets in boundary value problems. The latter arise in the separation of variables procedure for partial differential equations which are covered in the Year II Mathematics course in the context of heat and mass transfer.
Recall that a matrix is symmetric iff it is equal to its transpose, i.e.
Definition of Hermitian conjugate
If is a matrix with complex entries, the Hermitian conjugate of , denoted by is the complex conjugate of all entries in the matrix transposed, i.e.
A complex matrix is called Hermitian if
In particular, all symmetric matrices are Hermitian. We also note that the Hermitian conjugate of the product of two matrices is the product of their conjugates taken in reverse order
which follows from the same identity for transpose matrices and the fact that the complex conjugate of a product is the product of complex conjugates, i.e.
The diagonalisation of symmetric matrices is particularly easy as all eigenvalues are real and corresponding eigenvectors are orthonormal (these are mutually orthogonal vectors of unit length). We recall here that the dot product of two real vectors can be written using column vector notation as,
For complex-valued vectors the generalisation of the dot product involves the Hermitian conjugate. For two complex-valued vectors and , the dot product is defined as,
Hermitian matrices are orthogonally diagonalisable
While the proofs given here are beyond the scope of the course, the results are important. The key message is that Hermitian and therefore symmetric matrices are orthogonally diagonalisable.
Theorem 10.1 For a symmetric matrix , we may diagonalise with respect to an orthonormal set of eigenvectors.
The first step is to understand that the eigenvalues and therefore the eigenvectors of a Hermitian matrix are real.
Theorem 10.2 For an Hermitian matrix all eigenvalues are real.
Proof Consider an eigenvector with a possibly complex eigenvalue, . We want to show that is real which follows from . Starting from,
and multiplying by the Hermitian conjugate of ,
Next, we take the complex conjugate on both sides,
Using the property (10.126), (10.130) becomes,
Since is Hermitian, then which renders (10.131) as,
Finally, subtracting Eq. (10.132) from (10.129) gives,
Since is an eigenvector (and therefore nonzero), which implies that and the eigenvalues are real.
The second step is to show that eigenvectors corresponding to distinct eigenvalues are orthogonal. Theorem 10.3 Eigenvectors of Hermitian matrices corresponding to different eigenvalues are orthogonal.
Proof Consider two eigenvectors and of a Hermitian matrix , corresponding to eigenvalues and where . We want to show that and are orthogonal. We start from,
We multiply the first equation by
and the second by
We take the Hermitian conjugate on Eq. (10.135),
With similar arguments as in Theorem 10.2, we manipulate Eq. (10.136) as,
Finally, subtracting Eq. (10.137) from (10.134) gives,
Note that from Theorem 10.2, we know that since the eigenvalues of a Hermitian matrix are real. Further, since then which implies orthogonality.
The only question left to ask is - how can we prove that we have orthogonal eigenvectors for eigenvectors with the same eigenvalue? The first step towards answering this question requires deeper knowledge than is presented in this course and we assume that for a given eigenvalue we have a full set of eigenvectors. Once this is assumed, we can easily show that any set of eigenvectors can be turned into an orthogonal set of eigenvectors. This is done through the Gram-Schmidt process. Example 10.13 Let be
Diagonalise the symmetric matrix with respect to its orthonormal eigenvectors.
Solution We first compute the eigenvalues from,
hence the eigenvalues are 1, 2, and 3. Next, we compute the eigenvectors. Starting with , we need to solve
which has solution and and so the eigenvector is ; the unit vector is
Similarly, we find the other two unit vectors corresponding to and , respectively, as,
Recall that the diagonalisation matrix is given by , where is made up of the three eigenvectors as follows,
Since is a matrix with orthonormal columns then it is orthogonal which we know from Subsec. 10.1.4 to satisfy . Further, is the diagonal matrix with its diagonal entries made up of the eigenvalues of . Hence, we have
The biggest advantage of symmetric matrices is that we do not need to find the inverse of the matrix of eigenvectors using Gauss-Jordan, but we can just take the transpose.
Exercises
-
Using the Cayley-Hamilton theorem find the inverse of
(a) ;
(b) .
-
Orthogonally diagonalise the following matrices:
(a) ;
(b) .