A linear operator T on a finite-dimensional vector space V is called diagonalizable if there is an ordered basis β for V such that [T]β is a diagonal matrix. A square matrix A is called diagonalizable if LA is diagonalizable.
Let T be a linear operator on a vector space V. A nonzero vectorv∈V is called an eigenvector of T if there exists a scalar λ such that T(v)=λv. The scalar λ is called the eigenvalue corresponding to the eigenvector v.
Let A be in Mn×n(F). A nonzero vector v∈Fn is called an eigenvector of A if v is an eigenvector of LA; that is, if Av=λv for some scalar λ. The scalar λ is called the eigenvalue of A corresponding to the eigenvector v.
Theorem 5.1. Let A∈Mn×n(F). Then a scalar λ is an eigenvalue of A if and only if
Let A∈Mn×n(F). The polynomial f(t)=det(A−tIn) is called the characteristic polynomial of A.
It is easily shown that similar matrices have the same characteristic polynomial.
Let T be a linear operator on an n-dimensional vector space V with ordered basis β. We define the characteristic polynomialf(t) of T to be the characteristic polynomial of A=[T]β. That is,
f(t)=det(A−tIn).
We often denote the characteristic polynomial of an operator T by f(t)=det(T−tI).
Theorem 5.2. Let A∈Mn×n(F).
(a)The characteristic polynomial of A is a polynomial of degree n with leading coefficient (−1)n.
(b)A has at most n distinct eigenvalues.
Theorem 5.3. Let T be a linear operator on a vector space V, and let λ be an eigenvalue of T. A vector v∈V is an eigenvector of T corresponding to λ if and only if v=0 and v∈N(T−λI).
Theorem 5.4. Let T be a linear operator on a vector space V, and let λ1,λ2,…,λk be distinct eigenvalues of T. If v1,v2,…,vk are eigenvectors of T such that λi corresponds to vi(1≤i≤k), then {v1,v2,…,vk} is linearly independent.
Corollay 1. Let T be a linear operator on an n-dimensional vector space V. If T has n distinct eigenvalues, then T is diagonalizable.
Definition. A polynomial f(t) in P(F)splits over F if there are scalars c,a1,…,an (not necessarily distinct) in F such that
f(t)=c(t−a1)(t−a2)⋯(t−an).
If f(t) is the characteristic polynomial of a linear operator or a matrix over a field F, then the statement that f(t)splits is understood to mean that it splits over F.
Theorem 5.5. The characteristic polynomial of any diagonalizable linear operator splits.
Definition. Let λ be an eigenvalue of a linear operator or matrix with characteristic polynomial f(t). The (algebraic) multiplicity of λ is the largest positive integer k for which
Let T be a linear operator on a vector space V, and let λ be an eigenvalue of T. Define
Eλ={x∈V:T(x)=λx}=N(T−λIV).
The set Eλ is called the eigenspace of T corresponding to the eigenvalue λ. Analogously, we define the eigenspace of a square matrix A to be the eigenspace of LA.
Theorem 5.6. Let T be a linear operator on a finite-dimensional vector space V, and let λ be an eigenvalue of T having multiplicity m. Then
1≤dim(Eλ)≤m.
Theorem 5.7. Let T be a linear operator on a vector space V, and let λ1,λ2,…,λk be distinct eigenvalues of T. For each i=1,2,…,k, let Si be a finite linearly independent subset of the eigenspace Eλi. Then
S=S1∪S2∪⋯∪Sk
is a linearly independent subset of V.
Theorem 5.8. Let T be a linear operator on a finite-dimensional vector space V such that the characteristic polynomial of T splits. Let λ1,λ2,…,λk be the distinct eigenvalues of T. Then:
(a)T is diagonalizable if and only if the multiplicity of λi is equal to dim(Eλi) for all i.
(b)If T is diagonalizable and βi is an ordered basis for Eλi for each i, then
β=β1∪β2∪⋯∪βk
is an ordered basis for V consisting of eigenvectors of T.
The system of differential equations is written in matrix form as
x′=Ax,
where x(t) is the vector of unknown functions and A is the coefficient matrix.
The main idea is to diagonalizeA. If
A=QDQ−1,
then substituting into the system gives
x′=QDQ−1x.
Define the new variable
y(t)=Q−1x(t),
which transforms the system into
y′=Dy.
Since D is diagonal, this gives three independent scalar differential equations, which are easy to solve. The solution to the original system is obtained by transforming back:
Let T be a linear operator on a vector space V. A subspace W of V is called a T-invariant subspace of V if T(W)⊆W, that is, if T(v)∈W for all v∈W.
Definition. Let T be a linear operator on a vector space V, and let x be a nonzero vector in V. The subspace
W=span({x,T(x),T2(x),…})
is called the T-cyclic subspace of Vgenerated byx. It is a simple matter to show that W is T-invariant.
In fact, W is the “smallest’’ T-invariant subspace of V containing x. That is, any T-invariant subspace of V containing x must also contain W.
Theorem 5.11. Let T be a linear operator on a finite-dimensional vector space V, and let W be a T-invariant subspace of V. Then the characteristic polynomial of TW divides the characteristic polynomial of T.
Theorem 5.12. Let T be a linear operator on a finite-dimensional vector space V, and let W denote the T-cyclic subspace of V generated by a nonzero vector v∈V. Let k=dim(W). Then:
Theorem 5.13 (Cayley–Hamilton). Let T be a linear operator on a finite-dimensional vector space V, and let f(t) be the characteristic polynomial of T. Then f(T)=T0, the zero transformation. That is, T “satisfies’’ its characteristic equation.
Corollary 1. Let A be an n×n matrix, and let f(t) be the characteristic polynomial of A. Then f(A)=O, the n×n zero matrix.
Theorem 5.14. Let A be an n×n matrix and f(x) be a polynomial such that f(A)=O (the zero matrix). Then, every eigenvalue λ of A must be a root of the scalar equation f(x)=0.
Definition. Let B1∈Mm×m(F), and let B2∈Mn×n(F). We define the direct sum of B1 and B2, denoted B1⊕B2, as the (m+n)×(m+n) matrix A such that