Skip to main content
Logo image

Section 5.5 Diagonalization

Our treatment of eigenvectors in Section 5.4 was motivated in part by the objective of finding particularly simple matrix representations [T]B of a linear transformation T:VV. The simplest situation we could hope for is that there is a choice of basis B for which [T]B is diagonal. We say that the basis B diagonalizes the transformation T in this case, and that T is diagonalizable. In this section we develop theoretical and computational tools for determining whether a linear transformation T is diagonalizable, and for finding a diagonalizing basis B when T is in fact diagonalizable.

Subsection 5.5.1 Diagonalizable transformations

Definition 5.5.1. Diagonalizable.

Let V be a finite-dimensional vector space. A linear transformation T:VV is diagonalizable if there exists an ordered basis B of V for which [T]B is a diagonal matrix. In this case, we say the basis B diagonalizes T.
An n×n matrix A is diagonalizable if the matrix transformation TA:RnRn is diagonalizable.
As was already laid out in Section 5.4 a matrix representation [T]B is diagonal if the elements of B are eigenvectors of T. According to Theorem 5.5.2, the converse is also true.
Let B=(v1,v2,,vn) be an ordered basis of V. The matrix [T]B will be diagonal if and only if for each 1jn the j-th column of A is of the form
(0,,λj,0,,0)=λjej
for some λj. By Definition 5.2.1 the j-th column of [T]B is the coordinate vector [T(vj)]B. Thus [T]B is diagonal if and only if for all 1jn we have [T(vj)]B=λjej for some λjR. Next, by definition of [v]B, we have
[T(vj]B=(0,,λj,0,,0)T(vj)=λjvj.
We conclude that [T]B is diagonal if and only if vj is an eigenvector of T for all 1jn. Furthermore, when this is the case, we see that the j-th diagonal entry of [T]B is the corresponding eigenvalue λj. This proves statements (1) and (2). Statement (3) follows from (1) and Definition 5.5.1.
The phrase “an ordered basis consisting of eigenvectors of T” is a bit of a mouthful. The definition below allows us to shorten this to simply “an eigenbasis of T”.

Definition 5.5.3. Eigenbasis.

Let T:VV be a linear transformation. An ordered basis B=(v1,v2,,vn) is an eigenbasis of T if vj is an eigenvector of T for all 1jn.

Example 5.5.4.

Let T=TA, where
A=15[3443].
We saw in Example 5.4.9 that v1=(1,2) and v2=(1,2) are eigenvectors of T with eigenvalues 1,1, respectively. It is clear that the two eigenvectors are linearly independent, and hence that B=(v1,v2) is an eigenbasis of T. It follows from Theorem 5.5.2 that T is diagonalizable, and that in fact
[T]B=[1001],
as one easily verifies.

Example 5.5.5.

Let T:R2R2 be rotation by π/4: i.e., T=TA, where
A=12[2222].
As discussed in Example 5.4.10, T has no eigenvectors whatsoever. It follows that there is no eigenbasis of T, and hence that T is not diagonalizable.

Example 5.5.6.

Let T=TA, where
A=[2102].
As is easily computed, λ=2 is the only eigenvalue of T, and W2=span{(1,0)}. It follows that any two eigenvectors v1 and v2 lie in the one-dimensional space W2, and hence are scalar multiples of one another. Thus we cannot find two linearly independent eigenvectors of T. We conclude that T does not have an eigenbasis, and hence is not diagonalizable.

Subsection 5.5.2 Linear independence of eigenvectors

Roughly put, Theorem 5.5.2 tells us that T is diagonalizable if it has “enough” eigenvectors: more precisely, if we can find a large enough collection of linearly independent eigenvectors. So when exactly can we do this? Our first examples were deceptively simple in this regard due to their low-dimensional setting. For transformations of higher-dimensional spaces we need more theory, which we now develop. Theorem 5.5.7 will serve as one of the key results for our purposes. It tells us that eigenvectors chosen from different eigenspaces are linearly independent.
We prove the result by contradiction. Suppose we can find a finite set of eigenvectors with distinct eigenvalues that is linearly dependent. It follows that we can find such a set of minimum cardinality. In other words, there is positive integer r satisfying the following properties: (i) we can find a linearly dependent set of r eigenvectors of T with distinct eigenvalues; (ii) for all k<r, any set of k eigenvectors of T with distinct eigenvalues is linearly independent 1 .
Now assume S={v1,v2,,vr} is a set of minimal cardinality satisfying T(vi)=λivi for all 1in and λiλj for all 1i<jn. First observe that we must have r>1: eigenvectors are nonzero by definition, and thus any set consisting of a single eigenvector is linearly independent. Next, since S is linearly dependent we have
(5.5.1)c1v1+c2v2++crvr=0,
where ci0 for some 1ir. After reordering, we may assume without loss of generality that c10. Next we apply T to both sides of (5.5.1):
(5.5.2)c1v1+c2v2+crvr=0T(c1v1+c2v2++crvr)=T(0)(5.5.3)c1T(v1)+c2T(v2)++crT(vr)=0(5.5.4)c1λ1v1+c2λ2v2++crλrvr=0.
From equation (5.5.1) and the equation in (5.5.4) we have
λr(c1v1+c2v2+crvr)(c1λ1v1+c2λ2v2+crλrvr)=0,
and hence
(5.5.5)c1(λrλ1)v1++cr1(λrλr1)+cr(λrλr)vr=0.
Since c10 and λ1λr, we have c1(λrλ1)0. Thus equation (5.5.5) implies that the set S={v1,v2,,vr1} is a linearly dependent set of eigenvectors of T with distinct eigenvalues, contradicting the minimality of r. This completes our proof by contradiction.
Let S={v1,v2,,vn} be a set eigenvectors of T with distinct eigenvalues. According to Theorem 5.5.7 the set S is linearly independent. Since |S|=n=dimV it follows that B=(v1,v2,,vn) is an eigenbasis for T and hence T is diagonalizable.

Example 5.5.9.

Let T=TA, where
A=[662813781610].
The characteristic polynomial of A is
p(t)=t33t24t+12=(t+2)(t2)(t3).
Since A has three distinct eigenvalues the linear transformation TA is diagonalizable. Indeed, any choice of eigenvectors v1,v2,v3 with v1W2,v2W2,v3W3 is guaranteed to be linearly independent, and hence gives rise to an eigenbasis B=(v1,v2,v3) of TA. For example the usual procedure allows us to easily find eigenvectors
v1=(1,2,2),v2=(1,1,1),v3=(2,1,0)
from the three eigenspaces. You can verify for yourself that these three vectors are indeed linearly independent.

Remark 5.5.10.

Let T:VV be a linear transformation, dimV=n. It cannot be stressed enough that having n distinct eigenvalues is a sufficient, but not necessary condition for T to be diagonalizable. In other words we have
T has distinct eigenvaluesT diagonalizable
but
T diagonalizable T has distinct eigenvalues.
A good counterexample to keep in mind is TI:RnRn, where I=In is the n×n identity matrix. The transformation is clearly diagonalizable since [T]B=I, where B=(e1,e2,,en) is the standard basis; and yet λ=1 is the only eigenvalue of T.
Theorem 5.5.7 makes no assumption about the dimension of V and can thus can be applied to linear transformations of infinite-dimensional spaces. The differential operator T(f)=f provides an interesting example.

Example 5.5.11.

Let V=C(R), and let T:VV be defined as T(f)=f. For each λR let fλ(x)=eλx. In Example 5.4.12 we saw that the functions fλ are eigenvectors of T with eigenvalue λ: i.e., T(fλ)=λfλ. It follows from Corollary 5.5.8 that for any distinct values λ1,λ2,,λr the set {eλ1x,eλ2x,,eλrx} is linearly independent, and thus that the (uncountably) infinite set S={eλx:λR}C(R) is linearly independent.
The next corollary is a useful strengthening of Theorem 5.5.7, and will be used to prove Theorem 5.5.13. Roughly speaking, it says that eigenspaces associated to distinct eigenvalues are “linearly independent”. Be careful: the phrase in quotes currently has no real meaning for us. We know what it means for vectors to be linearly independent, but not subspaces. However, it is a decent shorthand for the precise statement of Corollary 5.5.12.
Before proving the result, we point out one subtlety here: although the wiWλi for all i, we cannot assume that each wi is an eigenvector. Indeed, wi is an eigenvector in this case if and only if wi0. This observation guides the proof that follows.
To pick out the terms of (5.5.6) that are nonzero (if any), we define
J={j:wj0}={j1,j2,,jk}.
Assume by contradiction that J is nonempty: i.e., |J|=k1. In this case we would have
0=w1+w2+wr=wj1+wj2++wjk,
since wi=0 for all iJ. But then
wj1+wj2++wjk=0
would be a nontrivial linear combination of the eigenvectors wji equal to 0. Since the eigenvectors wji have distinct eigenvalues, this contradicts Theorem 5.5.7. Thus J={}. Equivalently, wi=0 for all 1ir, as desired.
At last we are ready to state and prove what will be our main tool for determining whether a linear transformation is diagonalizable.
We prove the two implications separately. In each we use the equivalence
T is diagonalizableT has an eigenbasis B,
proved in Theorem 5.5.2.
Assume T is diagonalizable. From Theorem 5.5.2, there is an eigenbasis B of T. After reordering we may assume that
B=(vλ1,1,,vλ1,n1Wλ1,vλ2,1,,vλ2,n2Wλ2,,vλr,1,,vλr,nrWλr),
where for each 1ir and each 1jni, the element vλi,j is an eigenvector with eigenvalue λi: i.e., vλi,jWλi. Observer that since B is a list of n vectors, we have
n=n1+n2++nr.
We claim that for all 1ir the set Sλi={vλi,1,,vλi,ni} is a basis of Wλi. The desired result follows in this case since
i=1rdimWλi=i=1r|Sλi|=i=1rni=n.
Proceeding then to the claim, observe that each set Sλi is linearly independent, since the underlying set of B is linearly independent. Thus it suffices to show that spanSλi=Wλi for all 1ir. To this end, fix an i with 1in and take any vWλi. Since B is a basis we can write
v=j=1n1c1,jvλ1,jwλ1++j=1nici,jvλi,jwλi+j=1nrcr,jvλr,jwλr=w1+w2++wr,
where for each 1kr we have
wk=i=1nkck,jvλk,jWλk.
Bringing v to the right-hand side of the equation above yields
0=w1+w2++(wiv)++wr.
Recall that vWλi, and thus wivW)λi. Since wkWλk for all ki, it follows from Corollary 5.5.12 that
w1=w2==(wiv)==wr=0.
Thus
v=wi=j=1nici,jvλi,j,
showing that vspanSλi, as desired.
Let ni=dimWλi for all 1ir . We assume that
n=dimWλ1+dimWλ2+dimWλr=n1+n2++nr.
For each 1in, let
Sλi={vλi,1,vλi,2,,vλi,ni}
be a basis of the eigenspace Wλi. We claim
B=(vλ1,1,,vλ1,n1Wλ1,vλ2,1,,vλ2,n2Wλ2,,vλr,1,,vλr,nrWλr)
is an eigenbasis of T. Since 0vλi,jWλi for all 1ir and 1jni, we see that B consists of eigenvectors of T. Since
n1+n2+nr=n=dimV,
to show that B is a basis it suffices to show that it is linearly independent. To this end, assume we have
0=j=1n1c1,jvλ1,jwλ1+j=1n2c2,jvλ2,jwλ2+j=1nrcr,jvλr,jwλr=w1+w2++wr,
where for each 1ir we have
wi=i=1nici,jvλi,jWλk.
By Corollary 5.5.12 we must have
0=wi=i=1nici,jvλi,j
for all i. Finally, since the set
Sλi={vλi,1,vλi,2,,vλi,ni}
is linearly independent for each i, we must have ci,j=0 for all 1ir and 1jni. This proves that B is linearly independent, hence a basis.
We now collect our various results about diagonalizability into one procedure that (a) decides whether a linear transformation T is diagonalizable, and (b) if it is, computes an eigenbasis for T. The procedure applies to any linear transformation of a finite-dimensional vector space, not just matrix transformations. As usual, the first step is to choose a matrix representation A=[T]B for T.
For the most part the validity of this procedure is a direct consequence of Theorem 5.5.2 and Theorem 5.5.13. However, there are two details that need to be pointed out.
  • That T is diagonalizable if and only if A=[T]B is diagonalizable follows from the fact that a basis of the λ-eigenspace of A to a basis of the λ-eigenspace of T using the coordinate vector transformation [v]B.
  • That the ordered list B described in Step 3 is in fact a basis is shown in the proof of Theorem 5.5.13.

Example 5.5.15.

Let T=TA, where
A=[211032003].
Decide whether T is diagonalizable. If yes, find an eigenbasis of T and compute the corresponding matrix representing T.
Solution.
Note first that A=[T]B where B is the standard basis of R3. (See Theorem 5.2.3.) Since A is upper triangular, we easily see that its characteristic polynomial is p(t)=(t1)(t3)2. Next we investigate the eigenspaces:
W2=null(2IA)=null[011012001],coordinatecoordinateW3=null(3IA)=null[111002000].
By inspection we see that both 2IA and 3IA have rank 2, and hence nullity 32=1 by the rank-nullity theorem. Thus both eigenspaces have dimension one, and we have dimW2+dimW3=1+1=2<3. We conclude that A, and hence TA, is not diagonalizable.
The diagonalizability examples in this text will focus largely on the special case of matrix transformations TA:RnRn. However, our conscience demands that we give at least one full example of a more abstract linear transformation.

Example 5.5.16. Transposition.

Let S:M22M22 be the linear transformation defined as S(A)=AT. Decide whether S is diagonalizable. If yes, find an eigenbasis for S and compute the corresponding matrix representing S.
Solution.
We saw in Example 5.4.24 that
[S]B=[1000001001000001],
where B=(E11,E12,E21,E22) is the standard ordered basis of M22. Furthermore, we saw that 1 and 1 are the distinct eigenvalues of A=[S]B, and that
S1={(1,0,0,0),(0,1,1,0),(0,0,0,1)},S1=span{(0,1,1,0)}
are bases of W1 and W1, respectively. It follows that dimW1+dimW1=3+1=4, that A is diagonalizable, and that
B=((1,0,0,0),(0,1,1,0),(0,0,0,1),(0,1,1,0))
is an eigenbasis of A. We conclude that S is diagonalizable, and we lift B via [v]B to the eigenbasis
B={[1000],[0110],[0001],[0110]}
of S. Lastly, we have
[S]B=[1000010000100001].

Video example: deciding if diagonalizable.

Figure 5.5.17. Video: deciding if diagonalizable

Subsection 5.5.3 Diagonalizable matrices

In this subsection we will focus on matrix transformations TA:RnRn. Recall (5.2.3) that in this situation we have A=[T]B where B is the standard basis of Rn. As such Procedure 5.5.14 boils down to steps (2)-(3), and the eigenbasis B of A found in (3) is itself an eigenbasis for T=TA. Letting D=[T]B the change of basis formula (5.3.20) yields
D=P1AP,
where P=PBB. Lastly, since B is the standard basis of Rn, the change of basis matrix PBB is obtained by placing the j-th element of B as the j-th column for all 1jn. We record these observations as a separate procedure specifically for matrix transformations.
The process of finding P and D satisfying (5.5.7) is called diagonalizing the matrix A; and we say that the matrix P diagonalizes A in this case. (Of course this is possible if and only if A is diagonalizable.)

Example 5.5.19.

The matrix
A=[1421339122534112245421222338]
has characteristic polynomial p(t)=t46t3+9t2+4t12. Decide whether A is diagonalizable. If yes, find an invertible matrix P and diagonal matrix D such that D=P1AP.
Solution.
To factor p(t), we first look for integer roots dividing the constant term 12: i.e., we test whether any of ±1,±2,±3,±4,±6,±12 are roots. Luckily, we see that 1 is a root of p(t). Doing polynomial division of p(t) by (t+1) yields
p(t)=(t+1)(t37t2+16t12)q(t).
Repeating this factoring technique on q(t), we see that q(2)=0, and thus can continue to factor:
p(t)=(t+1)(t37t2+16t12)=(t+1)(t2)(t25t+6)=(t+1)(t2)2(t3).
We conclude that the eigenvalues of A are 1, 2, and 3. We now compute bases for the corresponding eigenspaces. The bases below were obtained using Procedure 3.8.10. We omit the details of the Gaussian elimination performed in each case. (Check for yourself!)
W1=null[1521339122634112246421222337]=span{(1,1,1,1)}W2=null[1221339122334112243421222340]=span{(3,2,0,2),(1,1,2,1)}W1=null[1121339122234112242421222341]=span{(3,5,6,4)}.
We have ski Since
dimW1+dimW2+dimW3=1+2+1=4=dimR4,
we conclude that A is diagonalizable. Furthermore, we have D=P1AP, where
P=[1313121510261214],D=[1000020000200003].
Recall that two square matrices A and A are similar if A=P1AP for some invertible matrix P (5.3.27). From the foregoing discussion it follows that a matrix A is diagonalizable if and only if it is similar to a diagonal matrix.
According to Theorem 5.3.28 the matrix A is similar to a diagonal matrix D if and only if there is a linear transformation T:RnRn and ordered bases B,B of Rn such that [T]B=A and [T]B=D. By definition such a T would be diagonalizable, since [T]B=D is diagonal. Since T is diagonalizable if and only if A=[T]B is diagonalizable, we conclude that A is similar to a diagonal matrix D if and only if A is diagonalizable.
We know from Theorem 5.3.28 that similar matrices can be thought of as two matrix representations of the same overlying linear transformation T. As such similar matrices share many of the same algebraic properties, as Theorem 5.5.21 details.
Statement (1) follows by taking Q=P1.
Let pA(t) and pA(t) be the characteristic polynomials of A and A, repsectively. We have
pA(t)=det(tIA)=det(tIP1AP)(A=P1AP)=det(P1tIPP1AP)(algebra)=det(P1(tIA)P)(left/right dist.)=det(P1)det(tIA)det(P)(2.5.26)=(det(P))1det(P)det(tIA)=det(tIA)=pA(t).
This proves statement (2).
Statement (3) follows from (2) since the eigenvalues of a matrix are the real roots of its characteristic polynomial. Furthermore, by Theorem 5.4.25 the trace and determinant of a matrix are equal to the sum and product of the roots of its characteristic polynomial. Thus (4) also follows from (2).
The proofs of statements (5)-(6) are left as exercises.
A diagonalizable matrix is similar to a diagonal matrix (5.5.20) and similar matrices share many essential properties (5.3.28, 5.5.21) In this spirit, a good way of thinking about a diagonalizable matrix is that it is “as good as diagonal”.
In practical terms, if A is diagonalizable, then we have
(5.5.8)D=P1APA=PDP1
where D is diagonal. This allows us to answer questions about A by first answering the question for D and then use the equations in (5.5.8) to translate the results back to A. What makes this method effective is that algebraic questions involving diagonal matrices are easy to answer! Before getting to some illustrative examples, we need a few results about the operation AP1AP, which is called conjugation by P.
The proof is left as an exercise.

Example 5.5.24. Diagonalizable: matrix powers.

Assume D=P1AP, where D is diagonal. The normally difficult computation Ak can be accomplished by first computing Dk (easy) and then observing that
Ak=(PDP1)k=PDkP1(Theorem 5.5.23,(2)).
For example, the matrix
A=[1311]
is diagonalizable and satisfies D=P1AP, where
P=[3111],D=[2002].
It follows that for any kZ we have
Ak=PDkP1=P[2k00(2)k]P1=14[32k+(2)k32k3(2)k2k(2)k2k+3(2)k].

Example 5.5.25. Diagonalizable: matrix polynomials.

Assume D=P1AP, where D is a diagonal n×n matrix. Let [D]ii=di. Given any polynomial f(x)=anxn+an1xn1+a1x+a0, we have
f(A)=f(PDP1)=Pf(D)P1(Theorem 5.5.23,(3)).
Furthermore, since D is diagonal, it follows that f(D) is also diagonal, and in fact its diagonal entries are given by f(di). This gives us an easy method of computing arbitrary polynomials of the matrix A.
Consider again the matrix A (and P and D) from Example 5.5.24. Let f(x)=x24. Since f(2)=f(2)=0, it follows that f(D)=D24I=0. We conclude that
f(A)=A24I=Pf(D)P1=P0P1=0,
as you can check directly.

Example 5.5.26.

A square-root of an n×n matrix A is a matrix B such that B2=A. If A and A are similar matrices, satisfying A=P1AP, then A has a square-root if and only if A has a square-root. Indeed, if B satisfies B2=A, then C=P1BP satisfies
C2=(P1BP)2=P1B2P=P1AP=A.
Similarly, if C satisfies C2=A, then B=PCP1 satisfies
B2=(PCP1)2=PC2P1=PAP1=A.
As an example, the matrix
A=[0213]
satisfies D=P1AP, where
P=[2111],D=[1002].
Since
C=[1002]
is a square-root of D,
B=PCP1=[222221+21+22]
is a square-root of A.
So when exactly does a diagonal matrix D have a square-root? Clearly, it is sufficient that the diagonal entries di satisfy di0 for all i, as in the example above. Interestingly, this is not a necessary condition! Indeed, consider the following example:
[1001]=[0110]2.

Subsection 5.5.4 Algebraic and geometric multiplicity

We end this section with a deeper look at what the characteristic polynomial reveals about eigenspaces. To begin with, we first define the characteristic polynomial of a general linear transformation T:VV, where V is a finite-dimensional vector space.

Definition 5.5.27. Characteristic polynomial of a transformation.

Let T:VV be a linear transformation, where V is finite-dimensional. Let B be an ordered basis of V, and let A=[T]B. We define the characteristic polynomial of T to be the characteristic polynomial of A: i.e., the characteristic polynomial of T is
p(t)=det(tIA).

Remark 5.5.28.

For the characteristic polynomial of a linear transformation T:VV to be well-defined, it should not depend on the choice of basis. This is true thanks to Theorem 5.5.21 and Theorem 5.3.20. Indeed, given two choice of ordered bases B,B of V, the matrices A=[T]B and A=[T]B are similar (5.3.20), and thus their characteristic polynomials are equal (5.5.21,(2)).
Let T:VV be a linear transformation, where V is finite-dimensional. If λR is an eigenvalue of T, then we can factor the chacteristic polynomial p(t) of T as p(t)=(tλ)mq(t), where λ is not a root of q(t). As we will see, the exponent m is an upper bound for the dimension of Wλ. We call m the algebraic multiplicity of the eigenvalue λ.

Definition 5.5.29. Algebraic/geometric multiplicity.

Let T:VV be a linear transformation, where V is finite-dimensional, and let p(t) be the characteristic polynomial of T. Given an eigenvalue λR of T, we can factor p(t) as p(t)=(tλ)mq(t), where λ is not a root of the polynomial q(t): i.e., q(λ)0. We call m the geometric multiplicity of the eigenvalue λ, and we call dimWλ its geometric multiplicity. If m>1, we say λ is a repeated eigenvalue of T.
Since λ is an eigenvalue, we have Wλ{0}, and thus dimWλ1. Assume by contradiction that dimWλ>m. Let m=dimWλ, and let Sλ={v1,v2,,vm} be a basis for Wλ. We can extend Sλ to an ordered basis
B=(v1,,vm,vm+1,,vn)
of V. By definition, the characteristic polynomial of T is given my p(t)=det(tIA), where A=[T]Bw. Since v1,v2,,vm are λ-eigenvectors of T, the matrix A=[T]B is of the form
An easy proof by induction on m shows that for such a matrix A we have p(t)=det(tIA)=(tλ)mr(t) for some polynomial r(t). On the other hand, since λ has algebraic multiplicity m we have p(t)=(tλ)mq(t) for some polynomial q(t) with q(λ)=0. Setting these two expressions equal to one another we see that
(tλ)mr(t)=(tλ)mq(t),
or equivalently,
(tλ)mmr(t)=q(t).
Since m>m it follows that q(λ)=(λλ)mmr(λ)=0. Contradiction! We conclude that dimWλm, as desired.
Implication: (2)(1).
If (2) is true, then each λi is an eigenvalue of T and we have
i=1rdimWλi=i=1rmi=n,
by counting degrees in (5.5.9). It follows from Theorem 5.5.2 that T is diagonalizable.
Implication: (1)(2).
If T is diagonalizable, then there is an ordered basis B of V for which D=[T]B is diagonal. Letting di be the i-th diagonal element of D, we have
p(t)=det(tID)=(td1)(td2)(tdn).
This expression tells us that d1,d2,,dn are the roots of p(t), and hence that all roots are real since since diR for all 1in. On the other hand each λi is a root of p(t), and thus λiR for all 1ir. It follows that λ1,λ2,,λr are the distinct eigenvalues of T. By Theorem 5.5.13, since T is diagonalizable we must have
(5.5.10)i=1rdimWλi=n.
Since dimWλimi for all 1in (5.5.30), and since i=1rmi=n (counting degrees in (5.5.9)), for the equality (5.5.10) to hold we must have dimWλi=mi for all 1ir, as desired.
From Theorem 5.5.30 and Corollary 5.5.31 we can deduce a much finer picture of the eigenspaces of a linear transformation from its factored characteristic polynomial. This often reduces our workload when treating questions of diagonalizability, as the next examples illustrate.

Example 5.5.32.

The matrix
A=[2110420441144112]
has characteristic polynomial p(t)=(t1)(t+2)(t2)2. Decide whether A is diagonalizable.
Solution.
The eigenvalues of A are 1,2,2. Since the eigenvalues 1 and 2 both have algebraic multiplicity 1, we have by Theorem 5.5.30
1dimW1,dimW21,
and hence
dimW1=dimW2=1.
It follows that A is diagonalizable if and only if dimW2=2. We have W2=null(2IA), where
2IA=[0110400441144114].
This matrix clearly has rank 2 (the first two columns form a basis for its column space), and hence nullity 42=2. We conclude that A is diagonalizable.

Example 5.5.33.

The matrix
A=[1031011100210010]
has characterisic polynomial p(t)=(t1)2(t+1)2. Decide whether A is diagonalizable.
Solution.
The eigenvalues of A are 1 and 1, and each has algebraic multiplicity 2. Thus 1dimW1,dimW12, and A is diagonalizable if and only if
dimW1=dimW1=2.
By inspection we see that (1,0,0,0) and (0,1,0,0) are 1-eigenvectors, and thus we must have dimW1=2. Next we have W1=null(IA) where
IA=[2031021100110011].
It is not difficult to see (either using Gaussian elimination or inspection) that this matrix has rank 3, and hence nullity 1. We conclude that dimW1=1<2, and hence A is not diagonalizable.

Exercises 5.5.5 Exercises

Exercise Group.

For each matrix A use Procedure 5.5.14 to determine whether it is diagonalizable. If yes, then produce an invertible matrix P and diagonal matrix D satisfying D=P1AP. For the last matrix the characteristic polynomial p(t) is provided for convenience.
5.
A=[1996251191794]; p(t)=t34t2+5t2

6.

Let A=[abcd]. Show that A is diagonalizable if and only if either (ad)2+4bc>0 or A=aI (i.e., a=d and b=c=0).

Exercise Group.

For each exercise construct a 3×3 matrix A satisfying the given conditions. Begin by showing that the given A must be diagonalizable.
8.
A has eigenspaces W2=span{(1,0,1),(1,1,1) and W1=span{(1,0,1)}.
9.
Aw=w for all wW={(x,y,z):x+y+z=0}, Ax=0 for x=(1,1,1).

10.

Assume A is a 3×3 matrix with eigenvalues 0, 1, and 1.
  1. Show that A is diagonalizable. Provide an explicit diagonal matrix D that A is similar to.
  2. Prove that An=A for all odd integers n1.

13.

According to Theorem 5.5.21 if A and B are similar, then they have the same rank. Show that the converse is false by showing that the matrices
A=[1000],B=[0100]
have the same rank, but are not similar.

14.

According to Theorem 5.5.21 if A and B are similar, then they have the same characteristic polynomial. Show that the converse is false by showing that the matrices
A=[1101],B=[1001]
have the same characteristic polynomial, but are not similar.

17.

Each matrix A below has characteristic polynomial p(t)=t33t+2. Use Procedure 5.5.14 to decide whether A is diagonalizable. If yes, provide an inverible P and diagonal D satisfying D=P1AP.
  1. A=[503613604]
  2. A=[233334345]
That we can find a minimal r in this sense is plausible enough, but we are secretly using the well-ordering principle of the integers here.