I.1 Statement of the theorem
According to the Cayley-Hamilton theorem, every square matrix A satisfies its own characteristic equation (Volume 1, section 9.3.1). Let the characteristic polynomial of A be
(I.1)
When the determinant is fully expanded and terms in the same power of λ are collected, one obtains
(I.2)
For example, let
(I.3)
(I.4)
and
(I.5)
It will become clear at the end of Section I.3 why a matrix of integers provides an especially appropriate example.
The Cayley-Hamilton theorem asserts that
χ(A)= 0, (I.6)
where 0 is the zero matrix and
(I.7)
For the example in Eqs. (I.3–I.5),
(I.8)
Exercise I.1.2 verifies that χ(A)= 0 for this example.
Exercises for Section I.1
I.1.1 Verify Eq. (I.5).
I.1.2 Verify that χ(A)= 0, where A is defined in Eq. (I.3).
I.2 Elementary proof
Since the Cayley-Hamilton theorem is a fundamental result in linear algebra, it is useful to give two proofs, an elementary one and a more general one. The elementary proof is valid when the eigenvectors of A span the vector space on which A acts. The eigenvalue-eigenvector equation can be written in the form
AV = VΛ (I.9)
where V is the matrix of eigenvectors,
V =[v1 ,..., vn] (I.10)
and Λ is a diagonal matrix, the elements of which are the eigenvalues of A (Volume 1, Eq. (9.217)). Because this proof requires that the eigenvalues of A must belong to the same number field F to which the matrix elements of A belong, F must be algebraically closed, meaning that the roots of every polynomial equation with coefficients in F must belong to F. For the purposes of this proof, it is convenient to assume that F = C, the field of complex numbers.
Because we have assumed that the eigenvectors span the entire vector space Cn , the matrix V is non-singular. Therefore the inverse matrix V−1 exists. Multiplying both sides of Eq. (I.9) from the right with V−1 results in the equation
A = VΛV−1. (I.11)
Now
Ak = VΛk V−1, (I.12)
from which it follows that
(I.13)
since each eigenvalue λk is a root of the characteristic equation χ(λ) = 0. This establishes Eq. (I.6) for the case in which A is diagonalizable.
General proof
The general proof assumes only scalars that belong to a number field F, and n × n square matrices. A matrix of scalars is one in which every element is a scalar that belongs to F.A λ-matrix B(λ) is a matrix, the elements of which are polynomials over F in an unknown λ.A λ-matrix is equal to the zero matrix, 0, if and only if every matrix element is equal to the zero polynomial.
If, in each matrix element of a λ-matrix, one collects terms in powers of λ, the resulting λ-matrix is equal to a sum of matrices of scalars, Bj , each one multiplied by a power of λ:
(I.14)
where Bl ≠ 0 unless B(λ)= 0.
For example, let
Then
IatnhdatphpaetnBs that B(λ) is the matrix of cofactors of the matrix A − λ1 in Eq. (I.4), 0 is the matrix of cofactors of A; more of this later.
To prepare for the proof of the Cayley-Hamilton theorem, it is useful to show that, if a λ-matrix of the form
C(λ)= B(λ)(A − λ1) (I.17)
is equal to a matrix of scalars, then C(λ)= 0. To see this, expand the right-hand side using Eq. (I.14):
(I.18)
Because C(λ) is equal to a matrix of scalars, the matrix coefficient of the leading term, Bl, must vanish, along with each of the matrix coefficients in the terms j = l, l − 1,..., 1. But Bl = 0 and Bl A − Bl−1 = 0 imply that Bl−1 = 0.
Continuing the chain of equalities downward in j , one sees that Bj = 0 for j = l, l − 1,..., 0. It follows that C(λ)= 0.
Let B(λ) be the matrix of cofactors of the n x n matrix A − λ1. Each element of the cofactor matrix B(λ) is obtained from the matrix elements of A − λ1 by evaluating the determinant of a sub-matrix of A − λ1, and is therefore a polynomial of degree n − 1 or lower. Therefore B(λ) is a A-matrix as defined above:
(I.19)
By the Laplace expansion of a determinant (Volume 1, Eq. (6.347)),
(I.20)
where χ is the characteristic polynomial of the matrix A.
We now evaluate χ(A) and show that it is equal to the zero matrix. For every j ∈ (2 : n),
Aj − λj 1 = (Aj−1 + λAj−2 + ··· + λj−2A + λj−1) (A − λ1). (I.21)
Then
where G(λ)isa λ-matrix. Then
χ(A)= χ(λ)1 − G(λ)(A − λ1). (I.23)
Eq. (I.20) provides another expression for χ(λ)1. Substituting in Eq. (I.23), one obtains
χ(A) = B(λ) − G(λ)(A − λ1). (I.24)
But χ(A) is a matrix of scalars, and this equation is of the form of Eq. (I.17).
Therefore
χ(A)= 0. (I.25)
After a little study, one realizes that this proof makes no use of division by a scalar. In fact, the Cayley-Hamilton remains true if the number field F is replaced by a commutative ring R, such as the ring of integers, Z. In this case, the underlying vector space is replaced by an R-module (Volume 1, p. 207).
For example, if R = Z, then an underlying space of n × 1 column vectors with elements in a field F is replaced by a space whose elements are vectors of integers. The only allowable matrices that act on an R-module are matrices whose elements belong to R, as in Eq. (I.3), or are polynomials over R,as in Eq. (I.4). Determinants can still be defined. Each element of the matrix of cofactors is a determinant of a matrix of integers, and therefore is an integer, as in Eq. (I.15). The Laplace expansion of a determinant, Eq. (I.20), is still valid. Since every step of the proof remains valid for R-modules, the conclusion, Eq. (I.25), is also valid when the elements of A belong to a commutative ring R. The goal of Exercise I.3.2 is to verify the Cayley-Hamilton theorem for the matrix of integers defined in Eq. (I.3).
556 videos|198 docs
|
1. What is Cayley Hamilton theorem in matrix algebra? |
2. How is the Cayley Hamilton theorem used in matrix algebra? |
3. What is the significance of the Cayley Hamilton theorem in CSIR-NET Mathematical Sciences exam? |
4. Can the Cayley Hamilton theorem be applied to non-square matrices? |
5. Are there any limitations or exceptions to the Cayley Hamilton theorem? |
556 videos|198 docs
|
|
Explore Courses for Mathematics exam
|