Coordinate vectors and matrix representations work in tandem to model vectors in an abstract vector space \(V\) as column vectors in \(\R^n\text{,}\) and linear transformations \(T\colon V\rightarrow W\) as \(m\times n\) matrices. In both cases the model depends on our choice of basis. In this section we investigate how different basis choices affect these various models. Specifically, we consider the two questions below.
Given \(V\) and two ordered bases \(B\) and \(B'\text{,}\) what is the algebraic relation between \([\boldv]_B\) and \([\boldv]_{B'}\text{?}\)
Given \(T\colon V\rightarrow V\) and two ordered bases \(B\) and \(B'\text{,}\) what is the relation between \([T]_{B}\) and \([T]_{B'}\text{?}\)
We will tackle each question in turn. Both answers rely on something called a change of basis matrix\(\underset{B\rightarrow B'}{P}\text{.}\)
Subsection5.3.1Change of basis matrix
We define change of basis matrices via a column-by-column formula and motivate the definition retroactively with Theorem 5.3.2.
Definition5.3.1.Change of basis matrix.
Let \(B=(\boldv_1, \boldv_2, \dots, \boldv_n)\) and \(B'\) be two ordered bases for the vector space \(V\text{.}\) The change of basis from \(B\) to \(B'\) is the \(n\times n\) matrix \(\underset{B\rightarrow B'}{P}\) defined as
In other words, the \(j\)-th column of \(\underset{B\rightarrow B'}{P}\) is obtained by computing the coordinate vector of the \(j\)-th element of the original basis \(B\) with respect to the new basis \(B'\text{.}\)
Theorem5.3.2.Change of basis for coordinate vectors.
Let \(B\) and \(B'\) be two ordered bases of the \(n\)-dimensional vector space \(V\text{.}\)
Recall that \(\id_V\colon V\rightarrow V\) is the identity transformation (Definition 3.2.3), defined as \(\id_V(\boldv)=\boldv\) for all \(\boldv\in V\text{.}\) We have
In other words, to convert the \(B\)-coordinates of a vector \(\boldv\in V\) to \(B'\)-coordinates, simply multiply on the left by the matrix \(\underset{B\rightarrow B'}{P}\text{.}\)
Property (5.3.1) defines \(\underset{B\rightarrow B'}{P}\) uniquely: i.e., if \(A\) satisfies \(A[\boldv]_B=[\boldv]_{B'}\) for all \(\boldv\in \R^n\text{,}\) then \(A=\underset{B\rightarrow B'}{P}\text{.}\)
Let \(I_V\colon V\rightarrow V\) be the identity transformation: i.e., \(I_V(\boldv)=\boldv\) for all \(\boldv\in V\text{.}\) By Theorem 5.2.6 the matrix \([I_V]_B^{B'}\) is the unique matrix satisfying
By (2) of Theorem 5.2.6 (the uniqueness claim), if \(A\) satisfies \(A[\boldv]_B=[\boldv]_{B'}\) for all \(\boldv\in \R^n\text{,}\) then \(A=[\id_V]_{B}^{B'}\text{.}\) Since \([\id_V]_{B}^{B'}=\underset{B\rightarrow B'}{P}\text{,}\) we conclude \(A=\underset{B\rightarrow B'}{P}\text{.}\)
Example5.3.3.
Let \(V=\R^2\text{,}\)\(B=(\boldv_1=(1,1),\boldv_2=(1,-1))\text{,}\)\(B'=(\boldw_1=(1,2), \boldw_2=(2,-1))\text{.}\) Observe that \(B\) and \(B'\) are both orthogonal with resepct to the dot product. This will simplify your computations below.
Compute \(\underset{B\rightarrow B'}{P}\text{.}\)
Let \(\boldx=(4,-2)\text{.}\) Compute \([\boldx]_{B'}\) using (5.3.1).
The first two coordinate vector computations are nontrivial; you can verify for yourself that \(x^2=1(x-2)^2+4(x-2)+4\) and \(x=0(x-2)^2+1(x-2)+2\text{.}\) Alternatively, see Remark 5.3.5) for a neat trick for computing these coordinate vectors.
Since \(B\) is the standard basis, we see easily that \([x^2+x+1]_{B}=(1,1,1)\text{.}\) Using (5.3.1) we have
Let \(B=(x^n, x^{n-1}, \dots, x, 1)\) be the standard basis of \(P_n\text{.}\) Fix any constant \(a\in \R\text{,}\) and let \(B'=((x-a)^n, (x-a)^{n-1}, \dots, (x-a), 1)\text{.}\) It is easy to see that \(B'\) is also an ordered basis: a simple degree argument shows that the polynomials \(p_k(x)=(x-a)^k\) are linearly independent. It follows from Taylor’s theorem (from single-variable calculus) that given any polynomial \(p\in P_n\) we have
Let \(B=(\boldv_1, \boldv_2, \dots, \boldv_n)\text{.}\) By definition, the \(j\)-th column of \(\underset{B\rightarrow B'}{P}\) is \([\boldv_j]_B\text{.}\) Since \(\boldv_j=0\boldv_1+0\boldv_2+\cdots +1\boldv_j+\cdots\text{,}\) we see that \([\boldv_j]_B=\bolde_j\text{,}\) and hence that \(\underset{B\rightarrow B'}{P}=I\text{,}\) as claimed.
Let \(A=\underset{B\rightarrow B'}{P}\) and \(B=\underset{B'\rightarrow B"}{P}\text{.}\) For any \(\boldv\in V\) we have
It follows from (3) of Theorem 5.3.2 that \(BA=\underset{B\rightarrow B}{P}\text{.}\) Since \(\underset{B\rightarrow B}{P}=I\) by (1), we conclude that \(BA=I\text{,}\) and hence \(B=A^{-1}\text{,}\) as desired.
Example5.3.7.\(V=\R^n\text{,}\)\(B\) standard basis.
Consider the special situation where \(V=\R^n\text{,}\)\(B\) is the standard basis, and \(B'=\{\boldv_1,\dots,\boldv_n\}\) is some nonstandard basis. In this case we have
In other words, \(\underset{B'\rightarrow B}{P}\) is the matrix whose \(j\)-th column is just the \(j\)-th element of \(B'\text{.}\) Thus, in this situation we can compute \(\underset{B'\rightarrow B}{P}\) by placing the elements of \(B'\) as columns of a matrix, and then use (2) of Theorem 5.3.6 to compute \(\underset{B\rightarrow B'}{P}=\left(\underset{B'\rightarrow B}{P}\right)^{-1}\text{.}\)
Example5.3.8.
Let \(V=\R^2\text{,}\)\(B=((1,0),(0,1))\text{,}\)\(B'=\{(1,\sqrt{3}),(-\sqrt{3},1)\}\text{.}\) Compute \(\underset{B\rightarrow B'}{P}\) and \(\underset{B'\rightarrow B}{P}\text{.}\)
The observation from Example 5.3.7 applies more generally when \(B\) is the standard basis of the given vector space \(V\) and \(B'=(\boldv_1, \boldv_2, \dots, \boldv_n)\) is nonstandard. In this case computing \(\underset{B'\rightarrow B}{P}\) will be easy as the coordinate vectors \([\boldv_j]_{B}\) can be produced by inspection. See Example 5.3.10.
Example5.3.10.
Let \(V=M_{22}\text{,}\)\(B=(E_{11}, E_{12}, E_{21}, E_{22})\) (standard basis) and \(B'=(A_1,A_2,A_3,A_4)\text{,}\) where
Here the coordinate vectors \([A_i]_B\) are easily computed by inspection since \(B\) is the standard basis.
It turns out that \(\underset{B\rightarrow B'}{P}=(\underset{B'\rightarrow B}{P})^{-1}\) is not so difficult to compute in this case since the columns \(\boldc_j\) of \(\underset{B'\rightarrow B}{P}\) satisfy
Let \(B\) and \(B'\) be ordered bases of an \(n\)-dimensional inner product space \((V,\langle\, , \rangle)\text{.}\) Not surprisingly, if one or more of \(B\) and \(B'\) are orthornormal, our change of basis computations are simplified significantly. For example, if \(B'\) is orthonormal (or even orthogonal), the coordinate vector computations necessary to compute \(\underset{B\rightarrow B'}{P}\) can be done using the inner product formula of Theorem 5.1.9. Furthermore, as we see below, if both \(B\) and \(B'\) are orthonormal, then the columns of \(\underset{B\rightarrow B'}{P}\) form an orthonormal basis of \(\R^n\text{,}\) in which case we can compute \(\underset{B'\rightarrow B}{P}\) as
is orthogonal, as one easily checks by computing \(A^TA\text{.}\) Observe that the columns of \(A\) form an orthonormal set with respect to the dot product, as do the rows. This is not a coincidence!
Theorem5.3.15.Orthogonal matrices.
Let \(A\) be an \(n\times n\) matrix. The following statements are equivalent.
The matrix \(A\) is orthogonal.
The columns of \(A\) form an orthonormal basis of \(\R^n\) with respect to the dot product.
The rows of \(A\) form an orthonormal basis of \(\R^n\) with respect to the dot product.
Let \(\boldr_i\) and \(\boldc_i\) be the \(i\)-th row and column of \(A\text{,}\) respectively, for each \(1\leq i\leq n\text{.}\) From Theorem 4.1.9 we see that
We use here that rows of \(A^T\) are the columns of \(A\text{,}\) and the columns of \(A^T\) are the rows of \(A\text{.}\) From (5.3.3)–(5.3.4) it follows easily that
This proves \((1)\iff (2)\) and \((1)\iff (3)\text{.}\) The result follows.
Remark5.3.16.
It is somewhat unfortunate that the property of being an orthogonal matrix is equivalent to your rows or columns forming an orthonormal basis. You ask: Why not simply call such matrices orthonormal matrices? My answer: tradition!
Theorem5.3.17.Orthonormal change of basis.
Let \((V,\langle\, , \rangle)\) be a finite dimensional inner product space, and suppose \(B\) and \(B'\) are orthonormal bases of \(V\text{.}\)
The matrices \(\underset{B\rightarrow B'}{P}\) and \(\underset{B'\rightarrow B}{P}\) are orthogonal.
Let \(B=(\boldv_1, \boldv_2,\dots, \boldv_n)\text{.}\) By definition, the columns of \(\underset{B\rightarrow B'}{P}\) are the coordinate vectors \([\boldv_i]_{B'}\text{,}\)\(1\leq i\leq n\text{.}\) By Exercise 5.1.3.13, these coordinate vectors form an orthonormal subset of \(\R^n\text{;}\) since there are \(n=\dim\R^n\) of them, they form an orthonormal basis. From Theorem 5.3.15 it follows that \(\underset{B\rightarrow B'}{P}\) is orthogonal. Lastly, from Remark 5.3.13 it follows that \(\underset{B'\rightarrow B}{P}=\left(\underset{B\rightarrow B'}{P}\right)^{-1}\) is also orthogonal.
Since \(\underset{B\rightarrow B'}{P}\) is orthogonal, we have
Before connecting change of basis matrices with matrix representations of linear transformations, it is worth gathering some the different techniques for computing change of basis matrices we have discussed so far.
Procedure5.3.19.Change of basis computational tips.
Let \(B=(\boldv_1, \boldv_2, \dots, \boldv_n)\) and \(B'\) be ordered bases of the vector space \(V\text{.}\) Below you find a variety of techniques for computing \(\underset{B\rightarrow B'}{P}\) and \(\underset{B'\rightarrow B}{P}\text{.}\)
To compute \(\underset{B\rightarrow B'}{P}\) directly, we must compute \([\boldv_j]_{B'}\) for each \(1\leq j\leq n\text{.}\) This typically involves setting up and solving a linear system.
We have \(\underset{B'\rightarrow B}{P}=(\underset{B\rightarrow B'}{P})^{-1}\text{.}\) This observation is useful in situations where (a) one change of basis matrix is easier to compute than the other and (b) computing inverse matrices is not too onerous.
If \(B\) is the standard basis of \(V\text{,}\) then \(\underset{B'\rightarrow B}{P}\) is easy to compute. (See Remark 5.3.9.)
If \(B'\) is orthogonal with respect to some inner product on \(V\text{,}\) then we can easily compute \([\boldv_j]_{B'}\) for each \(1\leq j\leq n\) using Theorem 5.1.9.
If \(B\) and \(B'\) are both orthonormal bases of \(V\) with respect to a common inner product, then both change of basis matrices are orthogonal and we have \(\underset{B'\rightarrow B}{P}=(\underset{B\rightarrow B'}{P})^T\text{.}\)
Subsection5.3.3Change of basis for transformations
We now investigate how our choice of basis affects matrix representations of linear transformations. We will only consider the special case where \(T\colon V\rightarrow V\) and we are comparing matrix representations \([T]_B\) and \([T]_{B'}\) for two different ordered bases of \(V\text{.}\)
Theorem5.3.20.Change of basis for transformations.
Let \(V\) be finite-dimensional, let \(T\colon V\rightarrow V\) be linear, and let \(B\) and \(B'\) be two ordered bases for \(V\text{.}\) We have
First observe that (5.3.6) follows from (5.3.5) and (2) of Theorem 5.3.6. Next, to prove (5.3.5), it suffices by (2) of Theorem 5.2.6 to show that the matrix \(A=\underset{B\rightarrow B'}{P}\, [T]_B\, \underset{B'\rightarrow B}{P}\) satisfies
Remark5.3.21.Getting change of basis formulas correct.
It is easy to get the various details of the change of basis formula wrong. Here is a potential way to keep things organized in your mind.
We wish to relate \([T]_{B'}\) and \([T]_B\) with an equation of the form \([T]_{B'}=*[T]_B*\text{,}\) where the asterisks are to be replaced with change of basis matrices or their inverses. Think of the three matrices on the right-hand side of this equation as a sequence of three things done to coordinate vectors, reading from right to left.
\([T]_{B'}\) takes as inputs \(B'\)-coordinates of vectors, and outputs \(B'\)-coordinates. Thus the same should be true for \(*[T]_B*\text{.}\)
Since \([T]_B\) takes as inputs \(B\)-coordinates, we must first convert from \(B'\)-coordinates to \(B\)-coordinates. So we should have \([T]_{B'}=*[T]_B\underset{B'\rightarrow B}{P}\text{.}\)
Since \([T]_B\) outputs \(B\)-coordinates, we need to then convert back to \(B'\)-coordinates. Thus \([T]_{B'}=\underset{B\rightarrow B'}{P}[T]_B\underset{B'\rightarrow B}{P}\text{.}\)
If desired you may replace \(\underset{B\rightarrow B'}{P}\) with \(\underset{B'\rightarrow B}{P}^{-1}\text{.}\)
Example5.3.22.
Let \(T\colon P_2\rightarrow P_2\) be defined as \(T(p(x))=p(x)+2p'(x)+xp''(x)\text{.}\)
Let \(B=(x^2, x, 1)\text{.}\) Compute \([T]_B\text{.}\)
Let \(B'=(x^2+x+1, x^2+1, x+1)\text{.}\) Use the change of basis formula to compute \([T]_{B'}\text{.}\)
Consider the special case where \(T\colon \R^n\rightarrow \R^n\text{:}\) that is, when \(V=\R^n\) is a space of \(n\)-tuples. We know from Corollary 3.6.16 that \(T=T_A\) for a unique \(n\times n\) matrix \(A\text{.}\) Recall that \(A\) is called the standard matrix of \(T\) (3.6.17), and satisfies \(T(\boldx)=A\boldx\) for all \(\boldx\in \R^n\text{.}\) We often wish to compute \(A\text{,}\) as it provides a convenient matrix formula for \(T\text{.}\)
To compute \(A\)directly using the recipe in 3.6.16, we must compute \(T(\bolde_j)\) for each of the standard basis elements \(\bolde_j\text{.}\) For many naturally occurring transformations \(T\text{,}\) this is often not so easy to do. Theorem 5.3.20 provides an indirect method in such cases.
According to Theorem 5.2.3 we have \(A=[T]_B\text{:}\) i.e., the standard matrix of \(T\) is none other than the matrix representing \(T\) with respect to the standard basis. This connection allows us to compute \(A=[T]_B\) by first computing \([T]_{B'}\) for some more convenient basis \(B'\text{,}\) and then using the change of basis formula.
Procedure5.3.23.Computing the standard matrix using change of basis.
Let \(T\colon \R^n\rightarrow \R^n\) be a linear transformation, and let \(A\) be its standard matrix. To compute \(A\) using the change of basis formula (5.3.5), proceed as follows.
Find a convenient basis \(B'\) for which the action of \(T\) is easily understood.
Compute \(A'=[T]_{B'}\text{.}\)
Let \(B\) be the standard basis of \(\R^n\text{.}\) Recall that \(A=[T]_B\text{.}\) Now compute \(A\) using the change of basis formula as
Procedure 5.3.23 is a powerful technique for computing matrix formulas for many interesting geometric linear transformations of \(\R^n\text{:}\) e.g., rotations, reflections, and orthogonal projections. Often the very definition of such transformations will suggest a more convenient nonstandard basis \(B'\text{:}\) one that reflects the geometry involved. The next example illustrates this nicely.
Example5.3.24.Orthogonal projection (again).
Consider \(V=\R^3\) together with the dot product. Let’s derive (once again) a matrix formula for orthogonal projection \(\operatorname{proj}_W\colon \R^3\rightarrow \R^3\text{,}\) where \(W=\{(x,y,z)\colon x+y+z=0\}\text{.}\) In other words we want to compute \(A=[\operatorname{proj}_W]_B\text{,}\) where \(B=((1,0,0), (0,1,0), (0,0,1))\) is the standard basis. We will do so indirectly by first computing \([\operatorname{proj}_W]_{B'}\) with respect to a more convenient basis: namely, \(B'=((1,-1,0),(1,1,-2), (1,1,1))\text{.}\) This is the same basis from Example 5.2.12, and was selected deliberateley so that the first two vectors form a basis of \(W\text{,}\) and the third vector spans the normal line to \(W\text{.}\) As in Example 5.2.12 we then easily compute
Lo and behold, we’ve discovered our matrix formula for projection onto \(W\) once again! (Compare with Example 4.3.17 and Example 5.2.12.)
Video example: change of basis for transformations.
Video example: change of basis and reflection.
Subsection5.3.4Similarity and the holy commutative tent of linear algebra
Theorem 5.3.20 supplies an algebraic answer to the question: What is the relation between two matrix representations \(A=[T]_B\) and \(A'=[T]_{B'}\text{?}\) Letting \(P=\underset{B'\rightarrow B}{P}\text{,}\) equation (5.3.6) becomes \(A'=P^{-1}AP\text{.}\) Matrices satisfying such a relation are said to be similar.
Definition5.3.27.
Matrices \(A, A'\in M_{nn}\) are similar if there is an invertible matrix \(P\) such that \(A'=P^{-1}AP\text{.}\)
So any two matrix representations of a linear transformation \(T\colon V\rightarrow V\) are similar in the technical sense of Definition 5.3.27. In fact, a converse of sorts is also true, as articulated in the theorem below.
Theorem5.3.28.Similarity and matrix representations.
Two \(n\times n\) matrices \(A\) and \(A'\) are similar if and only if there is a linear transformation \(T\colon V\rightarrow V\) and bases \(B, B'\) of \(V\) satisfying \(A=[T]_B\) and \(A'=[T]_{B'}\text{.}\)
The discussion above shows that if \(A=[T]_B\) and \(A'=[T]_{B'}\text{,}\) then \(A'=P^{-1}AP\text{,}\) where \(P=\underset{B'\rightarrow B}{P}\text{;}\) thus \(A\) and \(A'\) are similar in this case.
Now assume that \(A\) and \(A'\) are similar. By definition this means there is an invertible matrix \(P\) such that \(A'=P^{-1}AP\text{.}\) Define \(T\colon \R^n\rightarrow \R^n\) as the matrix transformation \(T=T_A\text{.}\) According to Theorem 5.2.3 we have \(A=[T]_B\) where \(B\) is the standard basis of \(\R^n\text{.}\) Next, letting \(B'\) be the ordered basis whose \(j\)-th element is the \(j\)-th column of \(P\text{,}\) we have \(P=\underset{B'\rightarrow B}{P}\) (Example 5.3.7), and hence
We will see in Section 5.4 that similar matrices are indeed similar algebraically speaking: i.e., they share many of the same properties. Theorem 5.3.28 provides the theoretical foundation to understand why this should be so: if \(A\) and \(A'\) are similar, then they are two matrix representations of a common linear transformation \(T\text{;}\) their many shared properties are simply inherited from the single overlying linear transformation that they both represent! This circle of ideas is neatly encompassed by Figure 5.3.29.
Perhaps a little exegesis is in order here. Think of the map \(T\colon V\rightarrow V\) as a linear transformation up in abstract heaven; and think of the two matrices \(A=[T]_B\) and \(A'=[T]_{B'}\) as two earthly shadows of \(T\text{.}\) OK, this gets at the holy bit somewhat, but why commutative? Each face of the tent is a commutative diagram, as we now explain.
Slanted sides of the tent.
The commutativity of the two slanted sides of the tent is a consequence of Theorem 5.2.9:
Let \(P=\underset{B'\rightarrow B}{P}\text{,}\) so that \(P^{-1}=\underset{B\rightarrow B'}{P}\text{.}\) The commutativity of the two triangular ends of the tent are consequences of Theorem 5.3.2:
Since the base of the tent is commutative, and since the maps given by \(P\) and \(P^{-1}\) are invertible, we can translate back and forth between the matrices \(A\) and \(A'\text{.}\) Furthermore, since the two slanted sides of the tent are commutative, and since the coordinate vector transformations are invertible, we can translate up and down between our two matrix representations \(A\) and \(A'\) and the overlying linear transformation \(T\text{.}\) There is one true \(T\text{!}\)
Mantra5.3.30.Similar matrices mantra.
Similar matrices are but two shadows of a single overlying linear transformation.
Exercises5.3.5Exercises
Change of basis matrix.
In each exercise a vector space \(V\) is given along with two ordered bases \(B\) and \(B'\text{.}\) Compute \(\underset{B\rightarrow B'}{P}\) and \(\underset{B'\rightarrow B}{P}\text{.}\)
Compute \([x^2-x+4]_{B'}\) using the change of basis matrix \(\underset{B\rightarrow B'}{P}\) and (5.3.1).
7.
Let \(B\) be the standard basis of \(\R^2\text{.}\) Find the ordered basis \(B'\) for which the change of basis matrix \(\underset{B\rightarrow B'}{P}\) is given by
Let \(B\) be the standard basis of \(P_2\text{.}\) Find the ordered basis \(B'\) for which the change of basis matrix \(\underset{B'\rightarrow B}{P}\) is given by
In each exercise an inner product space \((V,\langle\, , \rangle)\) is given along with two orthonormal ordered bases \(B\) and \(B'\text{.}\) Compute \(\underset{B\rightarrow B'}{P}\) and \(\underset{B'\rightarrow B}{P}\text{.}\) Use Theorem 5.3.17 to reduce your workload.
9.
\(V=\R^2\) with the dot product, \(B=\left((\sqrt{2}/2, \sqrt{2}/2), (-\sqrt{2}/2, \sqrt{2}/2)\right)\text{,}\)\(B'=\left( (\sqrt{3}/2, -1/2), (1/2,\sqrt{3}/2)\right)\)
10.
\(V=\R^4\) with the dot product, \(B=(\bolde_1, \bolde_2, \bolde_3, \bolde_4)\text{,}\)\(B'=\left(\frac{1}{2}(1,1,1,1), \frac{1}{2}(1,-1,1,-1), \frac{1}{2}(1,1,-1,-1), \frac{1}{2}(1,-1,-1,1)\right)\)
11.
Suppose \(B=(\boldv_1, \boldv_2)\) and \(B'=(\boldw_1, \boldw_2)\) are two bases for the space \(V\) related by the change of basis matrix
To do so, set \(A=\underset{B'\rightarrow B''}{P}\) and \(B=\underset{B\rightarrow B'}{P}\) and show that the matrix \(AB\) satisfies the defining property of \(\underset{B\rightarrow B''}{P}\text{:}\) i.e.,
It is easy to verify that rotation and reflection matrices are orthogonal. We now know that conversely any orthogonal matrix is either a rotation or a reflection matrix.
17.
Let \(T\colon \R^3\rightarrow \R^3\) be the linear transformation defined as \(T(x,y,z)=(x+2y+z, -y, x+7z)\text{.}\) Let \(B\) be the standard basis of \(\R^3\text{,}\) and let \(B'=\left((1,0,0), (1,1,0), (1,1,1)\right)\text{.}\)
Let \(T\colon P_1\rightarrow P_1\) be the linear transformation defined as \(T(p(x))=(x+1)p'(x)\text{.}\) Let \(B\) be the standard basis of \(P_1\text{,}\) and let \(B'=\left(2x+1, x-1\right)\text{.}\)
Let \(\boldv=(a,b)\in \R^2\) be nonzero and define \(\ell=\Span\{\boldv\}\text{,}\) the line passing through the origin with direction vector \(\boldv\text{.}\) Let \(T\colon \R^2\rightarrow \R^2\) be reflection through \(\ell\text{.}\) (See Definition 3.2.16.) In this exercise we will use a change of basis argument to find a formula for the standard matrix of \(T\text{:}\) i.e., the matrix \(A\) satisfying \(T(\boldx)=A\boldx\) for all \(\boldx\in \R^2\text{.}\) Our answer will be expressed in terms of \(a\) and \(b\text{.}\)
Pick an orthonormal basis \(B'\) with one vector pointing along \(\ell\) and compute \([T]_{B'}\text{.}\)
Let \(B\) be the standard basis of \(\R^2\text{.}\) Use Theorem 5.3.20 to compute \(A=[T]_{B}\text{.}\) (Use Theorem 5.3.17 to reduce your workload.)
How do we know that \(A\) is the standard matrix of \(T\text{?}\)
Explain why your matrix \(A\text{,}\) expressed in terms of \(a\) and \(b\) for \(T\) agrees with the matrix formula provided in Theorem 3.2.17, which is expressed in terms of the angle \(\alpha\) that \(\ell\) makes with the \(x\)-axis.