Chapter 8 Operators on Complex Vector Spaces In this chapter we delve deeper into the structure of operators, with most of the attention on complex vector spaces. Some of the results in this chapter apply to both real and complex vector spaces; thus we do not make a standing assumption that ๐… = ๐‚ . Also, an inner product does not help with this material, so we return to the general setting of a finite-dimensional vector space. Even on a finite-dimensional complex vector space, an operator may not have enough eigenvectors to form a basis of the vector space. Thus we will consider the closely related objects called generalized eigenvectors. We will see that for each operator on a finite-dimensional complex vector space, there is a basis of the vector space consisting of generalized eigenvectors of the operator. The generalized eigenspace decomposition then provides a good description of arbitrary operators on a finite-dimensional complex vector space. Nilpotent operators, which are operators that when raised to some power equal 0 , have an important role in these investigations. Nilpotent operators provide a key tool in our proof that every invertible operator on a finite-dimensional complex vector space has a square root and in our approach to Jordan form. This chapter concludes by defining the trace and proving its key properties. standing assumptions for this chapter โ€ข ๐… denotes ๐‘ or ๐‚ . โ€ข ๐‘‰ denotes a finite-dimensional nonzero vector space over ๐… . D a v i d I li ff CC BY - SA The Long Room of the Old Library at the University of Dublin, where William Hamilton ( 1805โ€“1865 ) was a student and then a faculty member. Hamilton proved a special case of what we now call the Cayleyโ€“Hamilton theorem in 1853. Linear Algebra Done Right , fourth edition, by Sheldon Axler 297 Annotated Entity: ID: 311 Spans: True Boxes: True Text: 298 Chapter 8 Operators on Complex Vector Spaces 8A Generalized Eigenvectors and Nilpotent Operators Null Spaces of Powers of an Operator We begin this chapter with a study of null spaces of powers of an operator. 8.1 sequence of increasing null spaces Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then {0} = null ๐‘‡ 0 โІ null ๐‘‡ 1 โІ โ‹ฏ โІ null ๐‘‡ ๐‘˜ โІ null ๐‘‡ ๐‘˜ + 1 โІ โ‹ฏ . Proof Suppose ๐‘˜ is a nonnegative integer and ๐‘ฃ โˆˆ null ๐‘‡ ๐‘˜ . Then ๐‘‡ ๐‘˜ ๐‘ฃ = 0 , which implies that ๐‘‡ ๐‘˜ + 1 ๐‘ฃ = ๐‘‡(๐‘‡ ๐‘˜ ๐‘ฃ) = ๐‘‡(0) = 0 . Thus ๐‘ฃ โˆˆ null ๐‘‡ ๐‘˜ + 1 . Hence null ๐‘‡ ๐‘˜ โІ null ๐‘‡ ๐‘˜ + 1 , as desired. For similar results about decreasing sequences of ranges, see Exercises 6, 7, and 8. The following result states that if two consecutive terms in the sequence of sub- spaces above are equal, then all later terms in the sequence are equal. 8.2 equality in the sequence of null spaces Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘š is a nonnegative integer such that null ๐‘‡ ๐‘š = null ๐‘‡ ๐‘š + 1 . Then null ๐‘‡ ๐‘š = null ๐‘‡ ๐‘š + 1 = null ๐‘‡ ๐‘š + 2 = null ๐‘‡ ๐‘š + 3 = โ‹ฏ . Proof Let ๐‘˜ be a positive integer. We want to prove that null ๐‘‡ ๐‘š + ๐‘˜ = null ๐‘‡ ๐‘š + ๐‘˜ + 1 . We already know from 8.1 that null ๐‘‡ ๐‘š + ๐‘˜ โІ null ๐‘‡ ๐‘š + ๐‘˜ + 1 . To prove the inclusion in the other direction, suppose ๐‘ฃ โˆˆ null ๐‘‡ ๐‘š + ๐‘˜ + 1 . Then ๐‘‡ ๐‘š + 1 (๐‘‡ ๐‘˜ ๐‘ฃ) = ๐‘‡ ๐‘š + ๐‘˜ + 1 ๐‘ฃ = 0. Hence ๐‘‡ ๐‘˜ ๐‘ฃ โˆˆ null ๐‘‡ ๐‘š + 1 = null ๐‘‡ ๐‘š . Thus ๐‘‡ ๐‘š + ๐‘˜ ๐‘ฃ = ๐‘‡ ๐‘š (๐‘‡ ๐‘˜ ๐‘ฃ) = 0 , which means that ๐‘ฃ โˆˆ null ๐‘‡ ๐‘š + ๐‘˜ . This implies that null ๐‘‡ ๐‘š + ๐‘˜ + 1 โІ null ๐‘‡ ๐‘š + ๐‘˜ , completing the proof. The result above raises the question of whether there exists a nonnegative integer ๐‘š such that null ๐‘‡ ๐‘š = null ๐‘‡ ๐‘š + 1 . The next result shows that this equality holds at least when ๐‘š equals the dimension of the vector space on which ๐‘‡ operates. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 312 Spans: True Boxes: True Text: Section 8A Generalized Eigenvectors and Nilpotent Operators 299 8.3 null spaces stop growing Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then null ๐‘‡ dim ๐‘‰ = null ๐‘‡ dim ๐‘‰ + 1 = null ๐‘‡ dim ๐‘‰ + 2 = โ‹ฏ . Proof We only need to prove that null ๐‘‡ dim ๐‘‰ = null ๐‘‡ dim ๐‘‰ + 1 (by 8.2). Suppose this is not true. Then, by 8.1 and 8.2, we have {0} = null ๐‘‡ 0 โŠŠ null ๐‘‡ 1 โŠŠ โ‹ฏ โŠŠ null ๐‘‡ dim ๐‘‰ โŠŠ null ๐‘‡ dim ๐‘‰ + 1 , where the symbol โŠŠ means โ€œcontained in but not equal toโ€. At each of the strict inclusions in the chain above, the dimension increases by at least 1 . Thus dim null ๐‘‡ dim ๐‘‰ + 1 โ‰ฅ dim ๐‘‰ + 1 , a contradiction because a subspace of ๐‘‰ cannot have a larger dimension than dim ๐‘‰ . It is not true that ๐‘‰ = null ๐‘‡ โŠ• range ๐‘‡ for every ๐‘‡ โˆˆ โ„’ (๐‘‰) . However, the next result can be a useful substitute. 8.4 ๐‘‰ is the direct sum of null ๐‘‡ dim ๐‘‰ and range ๐‘‡ dim ๐‘‰ Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then ๐‘‰ = null ๐‘‡ dim ๐‘‰ โŠ• range ๐‘‡ dim ๐‘‰ . Proof Let ๐‘› = dim ๐‘‰ . First we show that 8.5 ( null ๐‘‡ ๐‘› ) โˆฉ ( range ๐‘‡ ๐‘› ) = {0}. Suppose ๐‘ฃ โˆˆ ( null ๐‘‡ ๐‘› ) โˆฉ ( range ๐‘‡ ๐‘› ) . Then ๐‘‡ ๐‘› ๐‘ฃ = 0 , and there exists ๐‘ข โˆˆ ๐‘‰ such that ๐‘ฃ = ๐‘‡ ๐‘› ๐‘ข . Applying ๐‘‡ ๐‘› to both sides of the last equation shows that ๐‘‡ ๐‘› ๐‘ฃ = ๐‘‡ 2๐‘› ๐‘ข . Hence ๐‘‡ 2๐‘› ๐‘ข = 0 , which implies that ๐‘‡ ๐‘› ๐‘ข = 0 (by 8.3). Thus ๐‘ฃ = ๐‘‡ ๐‘› ๐‘ข = 0 , completing the proof of 8.5. Now 8.5 implies that null ๐‘‡ ๐‘› + range ๐‘‡ ๐‘› is a direct sum (by 1.46). Also, dim ( null ๐‘‡ ๐‘› โŠ• range ๐‘‡ ๐‘› ) = dim null ๐‘‡ ๐‘› + dim range ๐‘‡ ๐‘› = dim ๐‘‰ , where the first equality above comes from 3.94 and the second equality comes from the fundamental theorem of linear maps (3.21). The equation above implies that null ๐‘‡ ๐‘› โŠ• range ๐‘‡ ๐‘› = ๐‘‰ (see 2.39), as desired. For an improvement of the result above, see Exercise 19. 8.6 example: ๐… 3 = null ๐‘‡ 3 โŠ• range ๐‘‡ 3 for ๐‘‡ โˆˆ โ„’ (๐… 3 ) Suppose ๐‘‡ โˆˆ โ„’ (๐… 3 ) is defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (4๐‘ง 2 , 0 , 5๐‘ง 3 ). Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 313 Spans: True Boxes: True Text: 300 Chapter 8 Operators on Complex Vector Spaces Then null ๐‘‡ = {(๐‘ง 1 , 0 , 0) โˆถ ๐‘ง 1 โˆˆ ๐…} and range ๐‘‡ = {(๐‘ง 1 , 0 , ๐‘ง 3 ) โˆถ ๐‘ง 1 , ๐‘ง 3 โˆˆ ๐…} . Thus null ๐‘‡ โˆฉ range ๐‘‡ โ‰  {0} . Hence null ๐‘‡ + range ๐‘‡ is not a direct sum. Also note that null ๐‘‡ + range ๐‘‡ โ‰  ๐… 3 . However, we have ๐‘‡ 3 (๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (0 , 0 , 125๐‘ง 3 ) . Thus we see that null ๐‘‡ 3 = {(๐‘ง 1 , ๐‘ง 2 , 0) โˆถ ๐‘ง 1 , ๐‘ง 2 โˆˆ ๐…} and range ๐‘‡ 3 = {(0 , 0 , ๐‘ง 3 ) โˆถ ๐‘ง 3 โˆˆ ๐…}. Hence ๐… 3 = null ๐‘‡ 3 โŠ• range ๐‘‡ 3 , as expected by 8.4. Generalized Eigenvectors Some operators do not have enough eigenvectors to lead to good descriptions of their behavior. Thus in this subsection we introduce the concept of generalized eigenvectors, which will play a major role in our description of the structure of an operator. To understand why we need more than eigenvectors, letโ€™s examine the question of describing an operator by decomposing its domain into invariant subspaces. Fix ๐‘‡ โˆˆ โ„’ (๐‘‰) . We seek to describe ๐‘‡ by finding a โ€œniceโ€ direct sum decomposition ๐‘‰ = ๐‘‰ 1 โŠ• โ‹ฏ โŠ• ๐‘‰ ๐‘š , where each ๐‘‰ ๐‘˜ is a subspace of ๐‘‰ invariant under ๐‘‡ . The simplest possible nonzero invariant subspaces are one-dimensional. A decomposition as above in which each ๐‘‰ ๐‘˜ is a one-dimensional subspace of ๐‘‰ invariant under ๐‘‡ is possible if and only if ๐‘‰ has a basis consisting of eigenvectors of ๐‘‡ (see 5.55). This happens if and only if ๐‘‰ has an eigenspace decomposition 8.7 ๐‘‰ = ๐ธ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐ธ(๐œ† ๐‘š , ๐‘‡) , where ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are the distinct eigenvalues of ๐‘‡ (see 5.55). The spectral theorem in the previous chapter shows that if ๐‘‰ is an inner product space, then a decomposition of the form 8.7 holds for every self-adjoint operator if ๐… = ๐‘ and for every normal operator if ๐… = ๐‚ because operators of those types have enough eigenvectors to form a basis of ๐‘‰ (see 7.29 and 7.31). However, a decomposition of the form 8.7 may not hold for more general operators, even on a complex vector space. An example was given by the operator in 5.57, which does not have enough eigenvectors for 8.7 to hold. Generalized eigenvectors and generalized eigenspaces, which we now introduce, will remedy this situation. 8.8 definition: generalized eigenvector Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† is an eigenvalue of ๐‘‡ . A vector ๐‘ฃ โˆˆ ๐‘‰ is called a generalized eigenvector of ๐‘‡ corresponding to ๐œ† if ๐‘ฃ โ‰  0 and (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘˜ ๐‘ฃ = 0 for some positive integer ๐‘˜ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 314 Spans: True Boxes: True Text: Section 8A Generalized Eigenvectors and Nilpotent Operators 301 Generalized eigenvalues are not de- fined because doing so would not lead to anything new. Reason: if (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘˜ is not injective for some positive inte- ger ๐‘˜ , then ๐‘‡ โˆ’ ๐œ†๐ผ is not injective, and hence ๐œ† is an eigenvalue of ๐‘‡ . A nonzero vector ๐‘ฃ โˆˆ ๐‘‰ is a general- ized eigenvector of ๐‘‡ corresponding to ๐œ† if and only if (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ ๐‘ฃ = 0 , as follows from applying 8.1 and 8.3 to the operator ๐‘‡ โˆ’ ๐œ†๐ผ . As we know, an operator on a complex vector space may not have enough eigenvectors to form a basis of the domain. The next result shows that on a complex vector space there are enough generalized eigenvectors to do this. 8.9 a basis of generalized eigenvectors Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then there is a basis of ๐‘‰ consisting of generalized eigenvectors of ๐‘‡ . Proof Let ๐‘› = dim ๐‘‰ . We will use induction on ๐‘› . To get started, note that the desired result holds if ๐‘› = 1 because then every nonzero vector in ๐‘‰ is an eigenvector of ๐‘‡ . This step is where we use the hypothesis that ๐… = ๐‚ , because if ๐… = ๐‘ then ๐‘‡ may not have any eigenvalues. Now suppose ๐‘› > 1 and the de- sired result holds for all smaller values of dim ๐‘‰ . Let ๐œ† be an eigenvalue of ๐‘‡ . Applying 8.4 to ๐‘‡ โˆ’ ๐œ†๐ผ shows that ๐‘‰ = null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› โŠ• range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› . If null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› = ๐‘‰ , then every nonzero vector in ๐‘‰ is a generalized eigen- vector of ๐‘‡ , and thus in this case there is a basis of ๐‘‰ consisting of generalized eigenvectors of ๐‘‡ . Hence we can assume that null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› โ‰  ๐‘‰ , which implies that range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› โ‰  {0} . Also, null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› โ‰  {0} , because ๐œ† is an eigenvalue of ๐‘‡ . Thus we have 0 < dim range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› < ๐‘›. Furthermore, range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› is invariant under ๐‘‡ [ by 5.18 with ๐‘(๐‘ง) = (๐‘ง โˆ’ ๐œ†) ๐‘› ] . Let ๐‘† โˆˆ โ„’ ( range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› ) equal ๐‘‡ restricted to range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› . Our induction hypothesis applied to the operator ๐‘† implies that there is a basis of range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› consisting of generalized eigenvectors of ๐‘† , which of course are generalized eigenvectors of ๐‘‡ . Adjoining that basis of range (๐‘‡โˆ’๐œ†๐ผ) ๐‘› to a basis of null (๐‘‡โˆ’๐œ†๐ผ) ๐‘› gives a basis of ๐‘‰ consisting of generalized eigenvectors of ๐‘‡ . If ๐… = ๐‘ and dim ๐‘‰ > 1 , then some operators on ๐‘‰ have the property that there exists a basis of ๐‘‰ consisting of generalized eigenvectors of the operator, and (unlike what happens when ๐… = ๐‚ ) other operators do not have this property. See Exercise 11 for a necessary and sufficient condition that determines whether an operator has this property. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 315 Spans: True Boxes: True Text: 302 Chapter 8 Operators on Complex Vector Spaces 8.10 example: generalized eigenvectors of an operator on ๐‚ 3 Define ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (4๐‘ง 2 , 0 , 5๐‘ง 3 ) for each (๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) โˆˆ ๐‚ 3 . A routine use of the definition of eigenvalue shows that the eigenvalues of ๐‘‡ are 0 and 5 . Furthermore, the eigenvectors corresponding to the eigenvalue 0 are the nonzero vectors of the form (๐‘ง 1 , 0 , 0) , and the eigenvectors corresponding to the eigenvalue 5 are the nonzero vectors of the form (0 , 0 , ๐‘ง 3 ) . Hence this operator does not have enough eigenvectors to span its domain ๐‚ 3 . We compute that ๐‘‡ 3 (๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (0 , 0 , 125๐‘ง 3 ) . Thus 8.1 and 8.3 imply that the generalized eigenvectors of ๐‘‡ corresponding to the eigenvalue 0 are the nonzero vectors of the form (๐‘ง 1 , ๐‘ง 2 , 0) . We also have (๐‘‡ โˆ’ 5๐ผ) 3 (๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (โˆ’125๐‘ง 1 + 300๐‘ง 2 , โˆ’125๐‘ง 2 , 0) . Thus the generalized eigenvectors of ๐‘‡ corresponding to the eigenvalue 5 are the nonzero vectors of the form (0 , 0 , ๐‘ง 3 ) . The paragraphs above show that each of the standard basis vectors of ๐‚ 3 is a generalized eigenvector of ๐‘‡ . Thus ๐‚ 3 indeed has a basis consisting of generalized eigenvectors of ๐‘‡ , as promised by 8.9. If ๐‘ฃ is an eigenvector of ๐‘‡ โˆˆ โ„’ (๐‘‰) , then the corresponding eigenvalue ๐œ† is uniquely determined by the equation ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ , which can be satisfied by only one ๐œ† โˆˆ ๐… (because ๐‘ฃ โ‰  0 ). However, if ๐‘ฃ is a generalized eigenvector of ๐‘‡ , then it is not obvious that the equation (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ ๐‘ฃ = 0 can be satisfied by only one ๐œ† โˆˆ ๐… . Fortunately, the next result tells us that all is well on this issue. 8.11 generalized eigenvector corresponds to a unique eigenvalue Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then each generalized eigenvector of ๐‘‡ corresponds to only one eigenvalue of ๐‘‡ . Proof Suppose ๐‘ฃ โˆˆ ๐‘‰ is a generalized eigenvector of ๐‘‡ corresponding to eigen- values ๐›ผ and ๐œ† of ๐‘‡ . Let ๐‘š be the smallest positive integer such that (๐‘‡โˆ’๐›ผ๐ผ) ๐‘š ๐‘ฃ = 0 . Let ๐‘› = dim ๐‘‰ . Then 0 = (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘› ๐‘ฃ = ((๐‘‡ โˆ’ ๐›ผ๐ผ) + (๐›ผ โˆ’ ๐œ†)๐ผ) ๐‘› ๐‘ฃ = ๐‘› โˆ‘ ๐‘˜=0 ๐‘ ๐‘˜ (๐›ผ โˆ’ ๐œ†) ๐‘›โˆ’๐‘˜ (๐‘‡ โˆ’ ๐›ผ๐ผ) ๐‘˜ ๐‘ฃ , where ๐‘ 0 = 1 and the values of the other binomial coefficients ๐‘ ๐‘˜ do not matter. Apply the operator (๐‘‡ โˆ’ ๐›ผ๐ผ) ๐‘šโˆ’1 to both sides of the equation above, getting 0 = (๐›ผ โˆ’ ๐œ†) ๐‘› (๐‘‡ โˆ’ ๐›ผ๐ผ) ๐‘šโˆ’1 ๐‘ฃ. Because (๐‘‡ โˆ’ ๐›ผ๐ผ) ๐‘šโˆ’1 ๐‘ฃ โ‰  0 , the equation above implies that ๐›ผ = ๐œ† , as desired. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 316 Spans: True Boxes: True Text: Section 8A Generalized Eigenvectors and Nilpotent Operators 303 We saw earlier (5.11) that eigenvectors corresponding to distinct eigenvalues are linearly independent. Now we prove a similar result for generalized eigen- vectors, with a proof that roughly follows the pattern of the proof of that earlier result. 8.12 linearly independent generalized eigenvectors Suppose that ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then every list of generalized eigenvectors of ๐‘‡ corresponding to distinct eigenvalues of ๐‘‡ is linearly independent. Proof Suppose the desired result is false. Then there exists a smallest positive integer ๐‘š such that there exists a linearly dependent list ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š of generalized eigenvectors of ๐‘‡ corresponding to distinct eigenvalues ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š of ๐‘‡ (note that ๐‘š โ‰ฅ 2 because a generalized eigenvector is, by definition, nonzero). Thus there exist ๐‘Ž 1 , โ€ฆ , ๐‘Ž ๐‘š โˆˆ ๐… , none of which are 0 (because of the minimality of ๐‘š ), such that ๐‘Ž 1 ๐‘ฃ 1 + โ‹ฏ + ๐‘Ž ๐‘š ๐‘ฃ ๐‘š = 0. Let ๐‘› = dim ๐‘‰ . Apply (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› to both sides of the equation above, getting 8.13 ๐‘Ž 1 (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ 1 + โ‹ฏ + ๐‘Ž ๐‘šโˆ’1 (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ ๐‘šโˆ’1 = 0. Suppose ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘š โˆ’ 1} . Then (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ ๐‘˜ โ‰  0 because otherwise ๐‘ฃ ๐‘˜ would be a generalized eigenvector of ๐‘‡ corresponding to the distinct eigenvalues ๐œ† ๐‘˜ and ๐œ† ๐‘š , which would contradict 8.11. However, (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) ๐‘› ((๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ ๐‘˜ ) = (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ((๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) ๐‘› ๐‘ฃ ๐‘˜ ) = 0. Thus the last two displayed equations show that (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ ๐‘˜ is a generalized eigenvector of ๐‘‡ corresponding to the eigenvalue ๐œ† ๐‘˜ . Hence (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ 1 , โ€ฆ , (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘› ๐‘ฃ ๐‘šโˆ’1 is a linearly dependent list (by 8.13) of ๐‘šโˆ’1 generalized eigenvectors correspond- ing to distinct eigenvalues, contradicting the minimality of ๐‘š . This contradiction completes the proof. Nilpotent Operators 8.14 definition: nilpotent An operator is called nilpotent if some power of it equals 0 . Thus an operator on ๐‘‰ is nilpotent if every nonzero vector in ๐‘‰ is a generalized eigenvector of ๐‘‡ corresponding to the eigenvalue 0 . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 317 Spans: True Boxes: True Text: 304 Chapter 8 Operators on Complex Vector Spaces 8.15 example: nilpotent operators (a) The operator ๐‘‡ โˆˆ โ„’ (๐… 4 ) defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 ) = (0 , 0 , ๐‘ง 1 , ๐‘ง 2 ) is nilpotent because ๐‘‡ 2 = 0 . (b) The operator on ๐… 3 whose matrix (with respect to the standard basis) is โŽ›โŽœโŽœโŽœโŽ โˆ’3 9 0 โˆ’7 9 6 4 0 โˆ’6 โŽžโŽŸโŽŸโŽŸโŽ  is nilpotent, as can be shown by cubing the matrix above to get the zero matrix. (c) The operator of differentiation on ๐’ซ ๐‘š (๐‘) is nilpotent because the (๐‘š + 1) th derivative of every polynomial of degree at most ๐‘š equals 0 . Note that on this space of dimension ๐‘š + 1 , we need to raise the nilpotent operator to the power ๐‘š + 1 to get the 0 operator. The Latin word nil means nothing or zero; the Latin word potens means having power. Thus nilpotent literally means having a power that is zero. The next result shows that when rais- ing a nilpotent operator to a power, we never need to use a power higher than the dimension of the space. For a slightly stronger result, see Exercise 18. 8.16 nilpotent operator raised to dimension of domain is 0 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Then ๐‘‡ dim ๐‘‰ = 0 . Proof Because ๐‘‡ is nilpotent, there exists a positive integer ๐‘˜ such that ๐‘‡ ๐‘˜ = 0 . Thus null ๐‘‡ ๐‘˜ = ๐‘‰ . Now 8.1 and 8.3 imply that null ๐‘‡ dim ๐‘‰ = ๐‘‰ . Thus ๐‘‡ dim ๐‘‰ = 0 . 8.17 eigenvalues of nilpotent operator Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . (a) If ๐‘‡ is nilpotent, then 0 is an eigenvalue of ๐‘‡ and ๐‘‡ has no other eigenvalues. (b) If ๐… = ๐‚ and 0 is the only eigenvalue of ๐‘‡ , then ๐‘‡ is nilpotent. Proof (a) To prove (a), suppose ๐‘‡ is nilpotent. Hence there is a positive integer ๐‘š such that ๐‘‡ ๐‘š = 0 . This implies that ๐‘‡ is not injective. Thus 0 is an eigenvalue of ๐‘‡ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 318 Spans: True Boxes: True Text: Section 8A Generalized Eigenvectors and Nilpotent Operators 305 To show that ๐‘‡ has no other eigenvalues, suppose ๐œ† is an eigenvalue of ๐‘‡ . Then there exists a nonzero vector ๐‘ฃ โˆˆ ๐‘‰ such that ๐œ†๐‘ฃ = ๐‘‡๐‘ฃ. Repeatedly applying ๐‘‡ to both sides of this equation shows that ๐œ† ๐‘š ๐‘ฃ = ๐‘‡ ๐‘š ๐‘ฃ = 0. Thus ๐œ† = 0 , as desired. (b) Suppose ๐… = ๐‚ and 0 is the only eigenvalue of ๐‘‡ . By 5.27(b), the minimal polynomial of ๐‘‡ equals ๐‘ง ๐‘š for some positive integer ๐‘š . Thus ๐‘‡ ๐‘š = 0 . Hence ๐‘‡ is nilpotent. Exercise 23 shows that the hypothesis that ๐… = ๐‚ cannot be deleted in (b) of the result above. Given an operator on ๐‘‰ , we want to find a basis of ๐‘‰ such that the matrix of the operator with respect to this basis is as simple as possible, meaning that the matrix contains many 0 โ€™s. The next result shows that if ๐‘‡ is nilpotent, then we can choose a basis of ๐‘‰ such that the matrix of ๐‘‡ with respect to this basis has more than half of its entries equal to 0 . Later in this chapter we will do even better. 8.18 minimal polynomial and upper-triangular matrix of nilpotent operator Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then the following are equivalent. (a) ๐‘‡ is nilpotent. (b) The minimal polynomial of ๐‘‡ is ๐‘ง ๐‘š for some positive integer ๐‘š . (c) There is a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ has the form โŽ›โŽœโŽœโŽœโŽ 0 โˆ— โ‹ฑ 0 0 โŽžโŽŸโŽŸโŽŸโŽ  , where all entries on and below the diagonal equal 0 . Proof Suppose (a) holds, so ๐‘‡ is nilpotent. Thus there exists a positive integer ๐‘› such that ๐‘‡ ๐‘› = 0 . Now 5.29 implies that ๐‘ง ๐‘› is a polynomial multiple of the minimal polynomial of ๐‘‡ . Thus the minimal polynomial of ๐‘‡ is ๐‘ง ๐‘š for some positive integer ๐‘š , proving that (a) implies (b). Now suppose (b) holds, so the minimal polynomial of ๐‘‡ is ๐‘ง ๐‘š for some positive integer ๐‘š . This implies, by 5.27(a), that 0 ( which is the only zero of ๐‘ง ๐‘š ) is the only eigenvalue of ๐‘‡ . This further implies, by 5.44, that there is a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ is upper triangular. This also implies, by 5.41, that all entries on the diagonal of this matrix are 0 , proving that (b) implies (c). Now suppose (c) holds. Then 5.40 implies that ๐‘‡ dim ๐‘‰ = 0 . Thus ๐‘‡ is nilpotent, proving that (c) implies (a). Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 319 Spans: True Boxes: True Text: 306 Chapter 8 Operators on Complex Vector Spaces Exercises 8A 1 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that if dim null ๐‘‡ 4 = 8 and dim null ๐‘‡ 6 = 9 , then dim null ๐‘‡ ๐‘š = 9 for all integers ๐‘š โ‰ฅ 5 . 2 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) , ๐‘š is a positive integer, ๐‘ฃ โˆˆ ๐‘‰ , and ๐‘‡ ๐‘šโˆ’1 ๐‘ฃ โ‰  0 but ๐‘‡ ๐‘š ๐‘ฃ = 0 . Prove that ๐‘ฃ , ๐‘‡๐‘ฃ , ๐‘‡ 2 ๐‘ฃ , โ€ฆ , ๐‘‡ ๐‘šโˆ’1 ๐‘ฃ is linearly independent. The result in this exercise is used in the proof of 8.45. 3 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‰ = null ๐‘‡ โŠ• range ๐‘‡ โŸบ null ๐‘‡ 2 = null ๐‘‡. 4 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) , ๐œ† โˆˆ ๐… , and ๐‘š is a positive integer such that the minimal polynomial of ๐‘‡ is a polynomial multiple of (๐‘ง โˆ’ ๐œ†) ๐‘š . Prove that dim null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š โ‰ฅ ๐‘š. 5 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘š is a positive integer. Prove that dim null ๐‘‡ ๐‘š โ‰ค ๐‘š dim null ๐‘‡. Hint: Exercise 21 in Section 3B may be useful. 6 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Show that ๐‘‰ = range ๐‘‡ 0 โЇ range ๐‘‡ 1 โЇ โ‹ฏ โЇ range ๐‘‡ ๐‘˜ โЇ range ๐‘‡ ๐‘˜ + 1 โЇ โ‹ฏ . 7 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘š is a nonnegative integer such that range ๐‘‡ ๐‘š = range ๐‘‡ ๐‘š + 1 . Prove that range ๐‘‡ ๐‘˜ = range ๐‘‡ ๐‘š for all ๐‘˜ > ๐‘š . 8 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that range ๐‘‡ dim ๐‘‰ = range ๐‘‡ dim ๐‘‰ + 1 = range ๐‘‡ dim ๐‘‰ + 2 = โ‹ฏ . 9 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘š is a nonnegative integer. Prove that null ๐‘‡ ๐‘š = null ๐‘‡ ๐‘š + 1 โŸบ range ๐‘‡ ๐‘š = range ๐‘‡ ๐‘š + 1 . 10 Define ๐‘‡ โˆˆ โ„’ (๐‚ 2 ) by ๐‘‡(๐‘ค , ๐‘ง) = (๐‘ง , 0) . Find all generalized eigenvectors of ๐‘‡ . 11 Suppose that ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that there is a basis of ๐‘‰ consisting of generalized eigenvectors of ๐‘‡ if and only if the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… . Assume ๐… = ๐‘ because the case ๐… = ๐‚ follows from 5.27 ( b ) and 8.9. This exercise states that the condition for there to be a basis of ๐‘‰ consisting of generalized eigenvectors of ๐‘‡ is the same as the condition for there to be a basis with respect to which ๐‘‡ has an upper-triangular matrix ( see 5.44 ) . Caution: If ๐‘‡ has an upper-triangular matrix with respect to a basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› of ๐‘‰ , then ๐‘ฃ 1 is an eigenvector of ๐‘‡ but it is not necessarily true that ๐‘ฃ 2 , โ€ฆ , ๐‘ฃ ๐‘› are generalized eigenvectors of ๐‘‡ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 320 Spans: True Boxes: True Text: Section 8A Generalized Eigenvectors and Nilpotent Operators 307 12 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that every vector in ๐‘‰ is a generalized eigenvector of ๐‘‡ . Prove that there exists ๐œ† โˆˆ ๐… such that ๐‘‡ โˆ’ ๐œ†๐ผ is nilpotent. 13 Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘†๐‘‡ is nilpotent. Prove that ๐‘‡๐‘† is nilpotent. 14 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent and ๐‘‡ โ‰  0 . Prove ๐‘‡ is not diagonalizable. 15 Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ is diagonalizable if and only if every generalized eigenvector of ๐‘‡ is an eigenvector of ๐‘‡ . For ๐… = ๐‚ , this exercise adds another equivalence to the list of conditions for diagonalizability in 5.55. 16 (a) Give an example of nilpotent operators ๐‘† , ๐‘‡ on the same vector space such that neither ๐‘† + ๐‘‡ nor ๐‘†๐‘‡ is nilpotent. (b) Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) are nilpotent and ๐‘†๐‘‡ = ๐‘‡๐‘† . Prove that ๐‘† + ๐‘‡ and ๐‘†๐‘‡ are nilpotent. 17 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent and ๐‘š is a positive integer such that ๐‘‡ ๐‘š = 0 . (a) Prove that ๐ผ โˆ’ ๐‘‡ is invertible and that (๐ผ โˆ’ ๐‘‡) โˆ’1 = ๐ผ + ๐‘‡ + โ‹ฏ + ๐‘‡ ๐‘šโˆ’1 . (b) Explain how you would guess the formula above. 18 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Prove that ๐‘‡ 1 + dimrange ๐‘‡ = 0 . If dim range ๐‘‡ < dim ๐‘‰ โˆ’ 1 , then this exercise improves 8.16. 19 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is not nilpotent. Show that ๐‘‰ = null ๐‘‡ dim ๐‘‰โˆ’1 โŠ• range ๐‘‡ dim ๐‘‰โˆ’1 . For operators that are not nilpotent, this exercise improves 8.4. 20 Suppose ๐‘‰ is an inner product space and ๐‘‡ โˆˆ โ„’ (๐‘‰) is normal and nilpotent. Prove that ๐‘‡ = 0 . 21 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that null ๐‘‡ dim ๐‘‰โˆ’1 โ‰  null ๐‘‡ dim ๐‘‰ . Prove that ๐‘‡ is nilpotent and that dim null ๐‘‡ ๐‘˜ = ๐‘˜ for every integer ๐‘˜ with 0 โ‰ค ๐‘˜ โ‰ค dim ๐‘‰ . 22 Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 5 ) is such that range ๐‘‡ 4 โ‰  range ๐‘‡ 5 . Prove that ๐‘‡ is nilpotent. 23 Give an example of an operator ๐‘‡ on a finite-dimensional real vector space such that 0 is the only eigenvalue of ๐‘‡ but ๐‘‡ is not nilpotent. This exercise shows that the implication ( b ) โŸน ( a ) in 8.17 does not hold without the hypothesis that ๐… = ๐‚ . 24 For each item in Example 8.15, find a basis of the domain vector space such that the matrix of the nilpotent operator with respect to that basis has the upper-triangular form promised by 8.18(c). 25 Suppose that ๐‘‰ is an inner product space and ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Show that there is an orthonormal basis of ๐‘‰ with respect to which the matrix of ๐‘‡ has the upper-triangular form promised by 8.18(c). Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 321 Spans: True Boxes: True Text: 308 Chapter 8 Operators on Complex Vector Spaces 8B Generalized Eigenspace Decomposition Generalized Eigenspaces 8.19 definition: generalized eigenspace, ๐บ(๐œ† , ๐‘‡) Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… . The generalized eigenspace of ๐‘‡ correspond- ing to ๐œ† , denoted by ๐บ(๐œ† , ๐‘‡) , is defined by ๐บ(๐œ† , ๐‘‡) = {๐‘ฃ โˆˆ ๐‘‰ โˆถ (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘˜ ๐‘ฃ = 0 for some positive integer ๐‘˜}. Thus ๐บ(๐œ† , ๐‘‡) is the set of generalized eigenvectors of ๐‘‡ corresponding to ๐œ† , along with the 0 vector. Because every eigenvector of ๐‘‡ is a generalized eigenvector of ๐‘‡ (take ๐‘˜ = 1 in the definition of generalized eigenvector), each eigenspace is contained in the corresponding generalized eigenspace. In other words, if ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… , then ๐ธ(๐œ† , ๐‘‡) โІ ๐บ(๐œ† , ๐‘‡) . The next result implies that if ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… , then the generalized eigenspace ๐บ(๐œ† , ๐‘‡) is a subspace of ๐‘‰ (because the null space of each linear map on ๐‘‰ is a subspace of ๐‘‰ ). 8.20 description of generalized eigenspaces Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… . Then ๐บ(๐œ† , ๐‘‡) = null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ . Proof Suppose ๐‘ฃ โˆˆ null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ . The definitions imply ๐‘ฃ โˆˆ ๐บ(๐œ† , ๐‘‡) . Thus ๐บ(๐œ† , ๐‘‡) โЇ null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ . Conversely, suppose ๐‘ฃ โˆˆ ๐บ(๐œ† , ๐‘‡) . Thus there is a positive integer ๐‘˜ such that ๐‘ฃ โˆˆ null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘˜ . From 8.1 and 8.3 (with ๐‘‡ โˆ’ ๐œ†๐ผ replacing ๐‘‡ ), we get ๐‘ฃ โˆˆ null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ . Thus ๐บ(๐œ† , ๐‘‡) โІ null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ , completing the proof. 8.21 example: generalized eigenspaces of an operator on ๐‚ 3 Define ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (4๐‘ง 2 , 0 , 5๐‘ง 3 ). In Example 8.10, we saw that the eigenvalues of ๐‘‡ are 0 and 5 , and we found the corresponding sets of generalized eigenvectors. Taking the union of those sets with {0} , we have ๐บ(0 , ๐‘‡) = {(๐‘ง 1 , ๐‘ง 2 , 0) โˆถ ๐‘ง 1 , ๐‘ง 2 โˆˆ ๐‚} and ๐บ(5 , ๐‘‡) = {(0 , 0 , ๐‘ง 3 ) โˆถ ๐‘ง 3 โˆˆ ๐‚}. Note that ๐‚ 3 = ๐บ(0 , ๐‘‡) โŠ• ๐บ(5 , ๐‘‡) . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 322 Spans: True Boxes: True Text: Section 8B Generalized Eigenspace Decomposition 309 In Example 8.21, the domain space ๐‚ 3 is the direct sum of the generalized eigenspaces of the operator ๐‘‡ in that example. Our next result shows that this behavior holds in general. Specifically, the following major result shows that if ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) , then ๐‘‰ is the direct sum of the generalized eigenspaces of ๐‘‡ , each of which is invariant under ๐‘‡ and on which ๐‘‡ is a nilpotent operator plus a scalar multiple of the identity. Thus the next result achieves our goal of decomposing ๐‘‰ into invariant subspaces on which ๐‘‡ has a known behavior. As we will see, the proof follows from putting together what we have learned about generalized eigenspaces and then using our result that for each operator ๐‘‡ โˆˆ โ„’ (๐‘‰) , there exists a basis of ๐‘‰ consisting of generalized eigenvectors of ๐‘‡ . 8.22 generalized eigenspace decomposition Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š be the distinct eigenvalues of ๐‘‡ . Then (a) ๐บ(๐œ† ๐‘˜ , ๐‘‡) is invariant under ๐‘‡ for each ๐‘˜ = 1 , โ€ฆ , ๐‘š ; (b) (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) is nilpotent for each ๐‘˜ = 1 , โ€ฆ , ๐‘š ; (c) ๐‘‰ = ๐บ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐บ(๐œ† ๐‘š , ๐‘‡) . Proof (a) Suppose ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘š} . Then 8.20 shows that ๐บ(๐œ† ๐‘˜ , ๐‘‡) = null (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) dim ๐‘‰ . Thus 5.18, with ๐‘(๐‘ง) = (๐‘งโˆ’๐œ† ๐‘˜ ) dim ๐‘‰ , implies that ๐บ(๐œ† ๐‘˜ , ๐‘‡) is invariant under ๐‘‡ , proving (a). (b) Suppose ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘š} . If ๐‘ฃ โˆˆ ๐บ(๐œ† ๐‘˜ , ๐‘‡) , then (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) dim ๐‘‰ ๐‘ฃ = 0 (by 8.20). Thus ((๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) ) dim ๐‘‰ = 0 . Hence (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) is nilpotent, proving (b). (c) To show that ๐บ(๐œ† 1 , ๐‘‡) + โ‹ฏ + ๐บ(๐œ† ๐‘š , ๐‘‡) is a direct sum, suppose ๐‘ฃ 1 + โ‹ฏ + ๐‘ฃ ๐‘š = 0 , where each ๐‘ฃ ๐‘˜ is in ๐บ(๐œ† ๐‘˜ , ๐‘‡) . Because generalized eigenvectors of ๐‘‡ cor- responding to distinct eigenvalues are linearly independent (by 8.12), this implies that each ๐‘ฃ ๐‘˜ equals 0 . Thus ๐บ(๐œ† 1 , ๐‘‡) + โ‹ฏ + ๐บ(๐œ† ๐‘š , ๐‘‡) is a direct sum (by 1.45). Finally, each vector in ๐‘‰ can be written as a finite sum of generalized eigen- vectors of ๐‘‡ (by 8.9). Thus ๐‘‰ = ๐บ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐บ(๐œ† ๐‘š , ๐‘‡) , proving (c). For the analogous result when ๐… = ๐‘ , see Exercise 8. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 323 Spans: True Boxes: True Text: 310 Chapter 8 Operators on Complex Vector Spaces Multiplicity of an Eigenvalue If ๐‘‰ is a complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) , then the decomposition of ๐‘‰ pro- vided by the generalized eigenspace decomposition (8.22) can be a powerful tool. The dimensions of the subspaces involved in this decomposition are sufficiently important to get a name, which is given in the next definition. 8.23 definition: multiplicity โ€ข Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . The multiplicity of an eigenvalue ๐œ† of ๐‘‡ is defined to be the dimension of the corresponding generalized eigenspace ๐บ(๐œ† , ๐‘‡) . โ€ข In other words, the multiplicity of an eigenvalue ๐œ† of ๐‘‡ equals dim null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ . The second bullet point above holds because ๐บ(๐œ† , ๐‘‡) = null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ (see 8.20). 8.24 example: multiplicity of each eigenvalue of an operator Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) is defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (6๐‘ง 1 + 3๐‘ง 2 + 4๐‘ง 3 , 6๐‘ง 2 + 2๐‘ง 3 , 7๐‘ง 3 ). The matrix of ๐‘‡ (with respect to the standard basis) is โŽ›โŽœโŽœโŽœโŽ 6 3 4 0 6 2 0 0 7 โŽžโŽŸโŽŸโŽŸโŽ  . The eigenvalues of ๐‘‡ are the diagonal entries 6 and 7 , as follows from 5.41. You can verify that the generalized eigenspaces of ๐‘‡ are as follows: ๐บ(6 , ๐‘‡) = span ((1 , 0 , 0) , (0 , 1 , 0)) and ๐บ(7 , ๐‘‡) = span ((10 , 2 , 1)). In this example, the multiplicity of each eigenvalue equals the number of times that eigenvalue appears on the diago- nal of an upper-triangular matrix rep- resenting the operator. This behavior always happens, as we will see in 8.31. Thus the eigenvalue 6 has multiplicity 2 and the eigenvalue 7 has multiplicity 1 . The direct sum ๐‚ 3 = ๐บ(6 , ๐‘‡) โŠ• ๐บ(7 , ๐‘‡) is the generalized eigenspace decom- position promised by 8.22. A basis of ๐‚ 3 consisting of generalized eigen- vectors of ๐‘‡ , as promised by 8.9, is (1 , 0 , 0) , (0 , 1 , 0) , (10 , 2 , 1) . There does not exist a basis of ๐‚ 3 consisting of eigen- vectors of this operator. In the example above, the sum of the multiplicities of the eigenvalues of ๐‘‡ equals 3 , which is the dimension of the domain of ๐‘‡ . The next result shows that this holds for all operators on finite-dimensional complex vector spaces. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 324 Spans: True Boxes: True Text: Section 8B Generalized Eigenspace Decomposition 311 8.25 sum of the multiplicities equals dim ๐‘‰ Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then the sum of the multiplicities of all eigenvalues of ๐‘‡ equals dim ๐‘‰ . Proof The desired result follows from the generalized eigenspace decomposition (8.22) and the formula for the dimension of a direct sum (see 3.94). The terms algebraic multiplicity and geometric multiplicity are used in some books. In case you encounter this terminology, be aware that the algebraic multi- plicity is the same as the multiplicity defined here and the geometric multiplicity is the dimension of the corresponding eigenspace. In other words, if ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† is an eigenvalue of ๐‘‡ , then algebraic multiplicity of ๐œ† = dim null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ = dim ๐บ(๐œ† , ๐‘‡) , geometric multiplicity of ๐œ† = dim null (๐‘‡ โˆ’ ๐œ†๐ผ) = dim ๐ธ(๐œ† , ๐‘‡). Note that as defined above, the algebraic multiplicity also has a geometric meaning as the dimension of a certain null space. The definition of multiplicity given here is cleaner than the traditional definition that involves determinants; 9.62 implies that these definitions are equivalent. If ๐‘‰ is an inner product space, ๐‘‡ โˆˆ โ„’ (๐‘‰) is normal, and ๐œ† is an eigenvalue of ๐‘‡ , then the algebraic multiplicity of ๐œ† equals the geometric multiplicity of ๐œ† , as can be seen from applying Exercise 27 in Section 7A to the normal operator ๐‘‡ โˆ’ ๐œ†๐ผ . As a special case, the singular values of ๐‘† โˆˆ โ„’ (๐‘‰ , ๐‘Š) (here ๐‘‰ and ๐‘Š are both finite-dimensional inner product spaces) depend on the multiplicities (either algebraic or geometric) of the eigenvalues of the self-adjoint operator ๐‘† โˆ— ๐‘† . The next definition associates a monic polynomial with each operator on a finite-dimensional complex vector space. 8.26 definition: characteristic polynomial Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š denote the distinct eigenvalues of ๐‘‡ , with multiplicities ๐‘‘ 1 , โ€ฆ , ๐‘‘ ๐‘š . The polynomial (๐‘ง โˆ’ ๐œ† 1 ) ๐‘‘ 1 โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) ๐‘‘ ๐‘š is called the characteristic polynomial of ๐‘‡ . 8.27 example: the characteristic polynomial of an operator Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) is defined as in Example 8.24. Because the eigenvalues of ๐‘‡ are 6 , with multiplicity 2 , and 7 , with multiplicity 1 , we see that the characteristic polynomial of ๐‘‡ is (๐‘ง โˆ’ 6) 2 (๐‘ง โˆ’ 7) . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 325 Spans: True Boxes: True Text: 312 Chapter 8 Operators on Complex Vector Spaces 8.28 degree and zeros of characteristic polynomial Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then (a) the characteristic polynomial of ๐‘‡ has degree dim ๐‘‰ ; (b) the zeros of the characteristic polynomial of ๐‘‡ are the eigenvalues of ๐‘‡ . Proof Our result about the sum of the multiplicities (8.25) implies (a). The definition of the characteristic polynomial implies (b). Most texts define the characteristic polynomial using determinants (the two definitions are equivalent by 9.62). The approach taken here, which is considerably simpler, leads to the following nice proof of the Cayleyโ€“Hamilton theorem. 8.29 Cayleyโ€“Hamilton theorem Suppose ๐… = ๐‚ , ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ž is the characteristic polynomial of ๐‘‡ . Then ๐‘ž(๐‘‡) = 0 . Proof Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š be the distinct eigenvalues of ๐‘‡ , and let ๐‘‘ ๐‘˜ = dim ๐บ(๐œ† ๐‘˜ , ๐‘‡) . For each ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘š} , we know that (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) is nilpotent. Thus we have Arthur Cayley ( 1821โ€“1895 ) published three mathematics papers before com- pleting his undergraduate degree. (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) ๐‘‘ ๐‘˜ | ๐บ(๐œ† ๐‘˜ , ๐‘‡) = 0 (by 8.16) for each ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘š} . The generalized eigenspace decom- position (8.22) states that every vector in ๐‘‰ is a sum of vectors in ๐บ(๐œ† 1 , ๐‘‡) , โ€ฆ , ๐บ(๐œ† ๐‘š , ๐‘‡) . Thus to prove that ๐‘ž(๐‘‡) = 0 , we only need to show that ๐‘ž(๐‘‡)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) = 0 for each ๐‘˜ . Fix ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘š} . We have ๐‘ž(๐‘‡) = (๐‘‡ โˆ’ ๐œ† 1 ๐ผ) ๐‘‘ 1 โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) ๐‘‘ ๐‘š . The operators on the right side of the equation above all commute, so we can move the factor (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) ๐‘‘ ๐‘˜ to be the last term in the expression on the right. Because (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ) ๐‘‘ ๐‘˜ | ๐บ(๐œ† ๐‘˜ , ๐‘‡) = 0 , we have ๐‘ž(๐‘‡)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) = 0 , as desired. The next result implies that if the minimal polynomial of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) has degree dim ๐‘‰ (as happens almost alwaysโ€”see the paragraphs following 5.24), then the characteristic polynomial of ๐‘‡ equals the minimal polynomial of ๐‘‡ . 8.30 characteristic polynomial is a multiple of minimal polynomial Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then the characteristic polynomial of ๐‘‡ is a polynomial multiple of the minimal polynomial of ๐‘‡ . Proof The desired result follows immediately from the Cayleyโ€“Hamilton theo- rem (8.29) and 5.29. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 326 Spans: True Boxes: True Text: Section 8B Generalized Eigenspace Decomposition 313 Now we can prove that the result suggested by Example 8.24 holds for all operators on finite-dimensional complex vector spaces. 8.31 multiplicity of an eigenvalue equals number of times on diagonal Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Suppose ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ such that โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) is upper triangular. Then the number of times that each eigenvalue ๐œ† of ๐‘‡ appears on the diagonal of โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) equals the multiplicity of ๐œ† as an eigenvalue of ๐‘‡ . Proof Let ๐ด = โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) . Thus ๐ด is an upper-triangular matrix. Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› denote the entries on the diagonal of ๐ด . Thus for each ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘›} , we have ๐‘‡๐‘ฃ ๐‘˜ = ๐‘ข ๐‘˜ + ๐œ† ๐‘˜ ๐‘ฃ ๐‘˜ for some ๐‘ข ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜โˆ’1 ) . Hence if ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘›} and ๐œ† ๐‘˜ โ‰  0 , then ๐‘‡๐‘ฃ ๐‘˜ is not a linear combination of ๐‘‡๐‘ฃ 1 , โ€ฆ , ๐‘‡๐‘ฃ ๐‘˜โˆ’1 . The linear dependence lemma (2.19) now implies that the list of those ๐‘‡๐‘ฃ ๐‘˜ such that ๐œ† ๐‘˜ โ‰  0 is linearly independent. Let ๐‘‘ denote the number of indices ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘›} such that ๐œ† ๐‘˜ = 0 . The conclusion of the previous paragraph implies that dim range ๐‘‡ โ‰ฅ ๐‘› โˆ’ ๐‘‘. Because ๐‘› = dim ๐‘‰ = dim null ๐‘‡ + dim range ๐‘‡ , the inequality above implies that 8.32 dim null ๐‘‡ โ‰ค ๐‘‘. The matrix of the operator ๐‘‡ ๐‘› with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is the upper- triangular matrix ๐ด ๐‘› , which has diagonal entries ๐œ† 1๐‘› , โ€ฆ , ๐œ† ๐‘›๐‘› [ see Exercise 2(b) in Section 5C ] . Because ๐œ† ๐‘˜๐‘› = 0 if and only if ๐œ† ๐‘˜ = 0 , the number of times that 0 appears on the diagonal of ๐ด ๐‘› equals ๐‘‘ . Thus applying 8.32 with ๐‘‡ replaced with ๐‘‡ ๐‘› , we have 8.33 dim null ๐‘‡ ๐‘› โ‰ค ๐‘‘. For ๐œ† an eigenvalue of ๐‘‡ , let ๐‘š ๐œ† denote the multiplicity of ๐œ† as an eigenvalue of ๐‘‡ and let ๐‘‘ ๐œ† denote the number of times that ๐œ† appears on the diagonal of ๐ด . Replacing ๐‘‡ in 8.33 with ๐‘‡ โˆ’ ๐œ†๐ผ , we see that 8.34 ๐‘š ๐œ† โ‰ค ๐‘‘ ๐œ† for each eigenvalue ๐œ† of ๐‘‡ . The sum of the multiplicities ๐‘š ๐œ† over all eigenvalues ๐œ† of ๐‘‡ equals ๐‘› , the dimension of ๐‘‰ (by 8.25). The sum of the numbers ๐‘‘ ๐œ† over all eigenvalues ๐œ† of ๐‘‡ also equals ๐‘› , because the diagonal of ๐ด has length ๐‘› . Thus summing both sides of 8.34 over all eigenvalues ๐œ† of ๐‘‡ produces an equality. Hence 8.34 must actually be an equality for each eigenvalue ๐œ† of ๐‘‡ . Thus the multiplicity of ๐œ† as an eigenvalue of ๐‘‡ equals the number of times that ๐œ† appears on the diagonal of ๐ด , as desired. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 327 Spans: True Boxes: True Text: โ‰ˆ 100๐œ‹ Chapter 8 Operators on Complex Vector Spaces Block Diagonal Matrices Often we can understand a matrix better by thinking of it as composed of smaller matrices. To interpret our results in matrix form, we make the following definition, gener- alizing the notion of a diagonal matrix. If each matrix ๐ด ๐‘˜ in the definition below is a 1 -by- 1 matrix, then we actually have a diagonal matrix. 8.35 definition: block diagonal matrix A block diagonal matrix is a square matrix of the form โŽ›โŽœโŽœโŽœโŽ ๐ด 1 0 โ‹ฑ 0 ๐ด ๐‘š โŽžโŽŸโŽŸโŽŸโŽ  , where ๐ด 1 , โ€ฆ , ๐ด ๐‘š are square matrices lying along the diagonal and all other entries of the matrix equal 0 . 8.36 example: a block diagonal matrix The 5 -by- 5 matrix ๐ด = โŽ›โŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœ โŽœ โŽœ โŽ ( 4 ) 0 0 0 0 0 0 โŽ›โŽœโŽ 2 โˆ’3 0 2 โŽžโŽŸโŽ  0 0 0 0 0 0 0 0 0 0 โŽ› โŽœ โŽ 1 7 0 1 โŽž โŽŸ โŽ  โŽžโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸ โŽŸ โŽŸ โŽ  is a block diagonal matrix with ๐ด = โŽ›โŽœโŽœโŽœโŽœโŽœโŽ ๐ด 1 0 ๐ด 2 0 ๐ด 3 โŽžโŽŸโŽŸโŽŸโŽŸโŽŸโŽ  , where ๐ด 1 = ( 4 ) , ๐ด 2 = โŽ›โŽœ โŽ 2 โˆ’3 0 2 โŽžโŽŸ โŽ  , ๐ด 3 = โŽ›โŽœ โŽ 1 7 0 1 โŽžโŽŸ โŽ  . Here the inner matrices in the 5 -by- 5 matrix above are blocked off to show how we can think of it as a block diagonal matrix. Note that in the example above, each of ๐ด 1 , ๐ด 2 , ๐ด 3 is an upper-triangular matrix whose diagonal entries are all equal. The next result shows that with respect to an appropriate basis, every operator on a finite-dimensional complex vector space has a matrix of this form. Note that this result gives us many more zeros in the matrix than are needed to make it upper triangular. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 328 Spans: True Boxes: True Text: Section 8B Generalized Eigenspace Decomposition 315 8.37 block diagonal matrix with upper-triangular blocks Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š be the distinct eigenvalues of ๐‘‡ , with multiplicities ๐‘‘ 1 , โ€ฆ , ๐‘‘ ๐‘š . Then there is a basis of ๐‘‰ with respect to which ๐‘‡ has a block diagonal matrix of the form โŽ›โŽœโŽœโŽœโŽ ๐ด 1 0 โ‹ฑ 0 ๐ด ๐‘š โŽžโŽŸโŽŸโŽŸโŽ  , where each ๐ด ๐‘˜ is a ๐‘‘ ๐‘˜ -by- ๐‘‘ ๐‘˜ upper-triangular matrix of the form ๐ด ๐‘˜ = โŽ›โŽœโŽœโŽœโŽ ๐œ† ๐‘˜ โˆ— โ‹ฑ 0 ๐œ† ๐‘˜ โŽžโŽŸโŽŸโŽŸโŽ . Proof Each (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) is nilpotent (see 8.22). For each ๐‘˜ , choose a basis of ๐บ(๐œ† ๐‘˜ , ๐‘‡) , which is a vector space of dimension ๐‘‘ ๐‘˜ , such that the matrix of (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) with respect to this basis is as in 8.18(c). Thus with respect to this basis, the matrix of ๐‘‡| ๐บ(๐œ† ๐‘˜ , ๐‘‡) , which equals (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) + ๐œ† ๐‘˜ ๐ผ| ๐บ(๐œ† ๐‘˜ , ๐‘‡) , looks like the desired form shown above for ๐ด ๐‘˜ . The generalized eigenspace decomposition (8.22) shows that putting together the bases of the ๐บ(๐œ† ๐‘˜ , ๐‘‡) โ€™s chosen above gives a basis of ๐‘‰ . The matrix of ๐‘‡ with respect to this basis has the desired form. 8.38 example: block diagonal matrix via generalized eigenvectors Let ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) be defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (6๐‘ง 1 + 3๐‘ง 2 + 4๐‘ง 3 , 6๐‘ง 2 + 2๐‘ง 3 , 7๐‘ง 3 ) . The matrix of ๐‘‡ (with respect to the standard basis) is โŽ›โŽœโŽœโŽœโŽ 6 3 4 0 6 2 0 0 7 โŽžโŽŸโŽŸโŽŸโŽ  , which is an upper-triangular matrix but is not of the form promised by 8.37. As we saw in Example 8.24, the eigenvalues of ๐‘‡ are 6 and 7 , and ๐บ(6 , ๐‘‡) = span ((1 , 0 , 0) , (0 , 1 , 0)) and ๐บ(7 , ๐‘‡) = span ((10 , 2 , 1)). We also saw that a basis of ๐‚ 3 consisting of generalized eigenvectors of ๐‘‡ is (1 , 0 , 0) , (0 , 1 , 0) , (10 , 2 , 1). The matrix of ๐‘‡ with respect to this basis is โŽ›โŽœโŽœโŽœ โŽ ( 6 3 0 6 ) 0 0 0 0 ( 7 ) โŽžโŽŸโŽŸโŽŸ โŽ  , which is a matrix of the block diagonal form promised by 8.37. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 329 Spans: True Boxes: True Text: 316 Chapter 8 Operators on Complex Vector Spaces Exercises 8B 1 Define ๐‘‡ โˆˆ โ„’ (๐‚ 2 ) by ๐‘‡(๐‘ค , ๐‘ง) = (โˆ’๐‘ง , ๐‘ค) . Find the generalized eigenspaces corresponding to the distinct eigenvalues of ๐‘‡ . 2 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible. Prove that ๐บ(๐œ† , ๐‘‡) = ๐บ( 1๐œ† , ๐‘‡ โˆ’1 ) for every ๐œ† โˆˆ ๐… with ๐œ† โ‰  0 . 3 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Suppose ๐‘† โˆˆ โ„’ (๐‘‰) is invertible. Prove that ๐‘‡ and ๐‘† โˆ’1 ๐‘‡๐‘† have the same eigenvalues with the same multiplicities. 4 Suppose dim ๐‘‰ โ‰ฅ 2 and ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that null ๐‘‡ dim ๐‘‰โˆ’2 โ‰  null ๐‘‡ dim ๐‘‰โˆ’1 . Prove that ๐‘‡ has at most two distinct eigenvalues. 5 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and 3 and 8 are eigenvalues of ๐‘‡ . Let ๐‘› = dim ๐‘‰ . Prove that ๐‘‰ = ( null ๐‘‡ ๐‘›โˆ’2 ) โŠ• ( range ๐‘‡ ๐‘›โˆ’2 ) . 6 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† is an eigenvalue of ๐‘‡ . Explain why the exponent of ๐‘ง โˆ’ ๐œ† in the factorization of the minimal polynomial of ๐‘‡ is the smallest positive integer ๐‘š such that (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š | ๐บ(๐œ† , ๐‘‡) = 0 . 7 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† is an eigenvalue of ๐‘‡ with multiplicity ๐‘‘ . Prove that ๐บ(๐œ† , ๐‘‡) = null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘‘ . If ๐‘‘ < dim ๐‘‰ , then this exercise improves 8.20. 8 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are the distinct eigenvalues of ๐‘‡ . Prove that ๐‘‰ = ๐บ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐บ(๐œ† ๐‘š , ๐‘‡) if and only if the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 ) ๐‘˜ 1 โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) ๐‘˜ ๐‘š for some positive integers ๐‘˜ 1 , โ€ฆ , ๐‘˜ ๐‘š . The case ๐… = ๐‚ follows immediately from 5.27 ( b ) and the generalized eigenspace decomposition ( 8.22 ) ; thus this exercise is interesting only when ๐… = ๐‘ . 9 Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that there exist ๐ท , ๐‘ โˆˆ โ„’ (๐‘‰) such that ๐‘‡ = ๐ท + ๐‘ , the operator ๐ท is diagonalizable, ๐‘ is nilpotent, and ๐ท๐‘ = ๐‘๐ท . 10 Suppose ๐‘‰ is a complex inner product space, ๐‘’ 1 , โ€ฆ , ๐‘’ ๐‘› is an orthonormal basis of ๐‘‡ , and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› be the eigenvalues of ๐‘‡ , each included as many times as its multiplicity. Prove that |๐œ† 1 | 2 + โ‹ฏ + |๐œ† ๐‘› | 2 โ‰ค โ€–๐‘‡๐‘’ 1 โ€– 2 + โ‹ฏ + โ€–๐‘‡๐‘’ ๐‘› โ€– 2 . See the comment after Exercise 5 in Section 7A. 11 Give an example of an operator on ๐‚ 4 whose characteristic polynomial equals (๐‘ง โˆ’ 7) 2 (๐‘ง โˆ’ 8) 2 . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 330 Spans: True Boxes: True Text: Section 8B Generalized Eigenspace Decomposition 317 12 Give an example of an operator on ๐‚ 4 whose characteristic polynomial equals (๐‘งโˆ’1)(๐‘งโˆ’5) 3 and whose minimal polynomial equals (๐‘งโˆ’1)(๐‘งโˆ’5) 2 . 13 Give an example of an operator on ๐‚ 4 whose characteristic and minimal polynomials both equal ๐‘ง(๐‘ง โˆ’ 1) 2 (๐‘ง โˆ’ 3) . 14 Give an example of an operator on ๐‚ 4 whose characteristic polynomial equals ๐‘ง(๐‘ง โˆ’ 1) 2 (๐‘ง โˆ’ 3) and whose minimal polynomial equals ๐‘ง(๐‘ง โˆ’ 1)(๐‘ง โˆ’ 3) . 15 Let ๐‘‡ be the operator on ๐‚ 4 defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 ) = (0 , ๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) . Find the characteristic polynomial and the minimal polynomial of ๐‘‡ . 16 Let ๐‘‡ be the operator on ๐‚ 6 defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 , ๐‘ง 5 , ๐‘ง 6 ) = (0 , ๐‘ง 1 , ๐‘ง 2 , 0 , ๐‘ง 4 , 0). Find the characteristic polynomial and the minimal polynomial of ๐‘‡ . 17 Suppose ๐… = ๐‚ and ๐‘ƒ โˆˆ โ„’ (๐‘‰) is such that ๐‘ƒ 2 = ๐‘ƒ . Prove that the characteris- tic polynomial of ๐‘ƒ is ๐‘ง ๐‘š (๐‘งโˆ’1) ๐‘› , where ๐‘š = dim null ๐‘ƒ and ๐‘› = dim range ๐‘ƒ . 18 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† is an eigenvalue of ๐‘‡ . Explain why the following four numbers equal each other. (a) The exponent of ๐‘ง โˆ’ ๐œ† in the factorization of the minimal polynomial of ๐‘‡ . (b) The smallest positive integer ๐‘š such that (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š | ๐บ(๐œ† , ๐‘‡) = 0 . (c) The smallest positive integer ๐‘š such that null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š = null (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š + 1 . (d) The smallest positive integer ๐‘š such that range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š = range (๐‘‡ โˆ’ ๐œ†๐ผ) ๐‘š + 1 . 19 Suppose ๐… = ๐‚ and ๐‘† โˆˆ โ„’ (๐‘‰) is a unitary operator. Prove that the constant term in the characteristic polynomial of ๐‘† has absolute value 1 . 20 Suppose that ๐… = ๐‚ and ๐‘‰ 1 , โ€ฆ , ๐‘‰ ๐‘š are nonzero subspaces of ๐‘‰ such that ๐‘‰ = ๐‘‰ 1 โŠ• โ‹ฏ โŠ• ๐‘‰ ๐‘š . Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and each ๐‘‰ ๐‘˜ is invariant under ๐‘‡ . For each ๐‘˜ , let ๐‘ ๐‘˜ denote the characteristic polynomial of ๐‘‡| ๐‘‰ ๐‘˜ . Prove that the characteristic polynomial of ๐‘‡ equals ๐‘ 1 โ‹ฏ๐‘ ๐‘š . 21 Suppose ๐‘ , ๐‘ž โˆˆ ๐’ซ (๐‚) are monic polynomials with the same zeros and ๐‘ž is a polynomial multiple of ๐‘ . Prove that there exists ๐‘‡ โˆˆ โ„’ (๐‚ deg ๐‘ž ) such that the characteristic polynomial of ๐‘‡ is ๐‘ž and the minimal polynomial of ๐‘‡ is ๐‘ . This exercise implies that every monic polynomial is the characteristic polynomial of some operator. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 331 Spans: True Boxes: True Text: 318 Chapter 8 Operators on Complex Vector Spaces 22 Suppose ๐ด and ๐ต are block diagonal matrices of the form ๐ด = โŽ›โŽœโŽœโŽœโŽ ๐ด 1 0 โ‹ฑ 0 ๐ด ๐‘š โŽžโŽŸโŽŸโŽŸโŽ  , ๐ต = โŽ›โŽœโŽœโŽœโŽ ๐ต 1 0 โ‹ฑ 0 ๐ต ๐‘š โŽžโŽŸโŽŸโŽŸโŽ  , where ๐ด ๐‘˜ and ๐ต ๐‘˜ are square matrices of the same size for each ๐‘˜ = 1 , โ€ฆ , ๐‘š . Show that ๐ด๐ต is a block diagonal matrix of the form ๐ด๐ต = โŽ›โŽœโŽœโŽœโŽ ๐ด 1 ๐ต 1 0 โ‹ฑ 0 ๐ด ๐‘š ๐ต ๐‘š โŽžโŽŸโŽŸโŽŸโŽ . 23 Suppose ๐… = ๐‘ , ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐œ† โˆˆ ๐‚ . (a) Show that ๐‘ข + ๐‘–๐‘ฃ โˆˆ ๐บ(๐œ† , ๐‘‡ ๐‚ ) if and only if ๐‘ข โˆ’ ๐‘–๐‘ฃ โˆˆ ๐บ(๐œ† , ๐‘‡ ๐‚ ) . (b) Show that the multiplicity of ๐œ† as an eigenvalue of ๐‘‡ ๐‚ equals the multiplicity of ๐œ† as an eigenvalue of ๐‘‡ ๐‚ . (c) Use (b) and the result about the sum of the multiplicities (8.25) to show that if dim ๐‘‰ is an odd number, then ๐‘‡ ๐‚ has a real eigenvalue. (d) Use (c) and the result about real eigenvalues of ๐‘‡ ๐‚ (Exercise 17 in Section 5A) to show that if dim ๐‘‰ is an odd number, then ๐‘‡ has an eigenvalue (thus giving an alternative proof of 5.34). See Exercise 33 in Section 3B for the definition of the complexification ๐‘‡ ๐‚ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 332 Spans: True Boxes: True Text: Section 8C Consequences of Generalized Eigenspace Decomposition 319 8C Consequences of Generalized Eigenspace Decomposition Square Roots of Operators Recall that a square root of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) is an operator ๐‘… โˆˆ โ„’ (๐‘‰) such that ๐‘… 2 = ๐‘‡ (see 7.36). Every complex number has a square root, but not every operator on a complex vector space has a square root. For example, the operator on ๐‚ 3 defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (๐‘ง 2 , ๐‘ง 3 , 0) does not have a square root, as you are asked to show in Exercise 1. The noninvertibility of that operator is no accident, as we will soon see. We begin by showing that the identity plus any nilpotent operator has a square root. 8.39 identity plus nilpotent has a square root Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Then ๐ผ + ๐‘‡ has a square root. Proof Consider the Taylor series for the function โˆš 1 + ๐‘ฅ : 8.40 โˆš 1 + ๐‘ฅ = 1 + ๐‘Ž 1 ๐‘ฅ + ๐‘Ž 2 ๐‘ฅ 2 + โ‹ฏ . Because ๐‘Ž 1 = 12 , the formula above implies that 1 + ๐‘ฅ2 is a good estimate for โˆš 1 + ๐‘ฅ when ๐‘ฅ is small. We do not find an explicit formula for the coefficients or worry about whether the infinite sum converges because we use this equation only as motivation. Because ๐‘‡ is nilpotent, ๐‘‡ ๐‘š = 0 for some positive integer ๐‘š . In 8.40, suppose we replace ๐‘ฅ with ๐‘‡ and 1 with ๐ผ . Then the infinite sum on the right side becomes a finite sum (because ๐‘‡ ๐‘˜ = 0 for all ๐‘˜ โ‰ฅ ๐‘š ). Thus we guess that there is a square root of ๐ผ + ๐‘‡ of the form ๐ผ + ๐‘Ž 1 ๐‘‡ + ๐‘Ž 2 ๐‘‡ 2 + โ‹ฏ + ๐‘Ž ๐‘šโˆ’1 ๐‘‡ ๐‘šโˆ’1 . Having made this guess, we can try to choose ๐‘Ž 1 , ๐‘Ž 2 , โ€ฆ , ๐‘Ž ๐‘šโˆ’1 such that the operator above has its square equal to ๐ผ + ๐‘‡ . Now (๐ผ + ๐‘Ž 1 ๐‘‡ + ๐‘Ž 2 ๐‘‡ 2 + ๐‘Ž 3 ๐‘‡ 3 + โ‹ฏ + ๐‘Ž ๐‘šโˆ’1 ๐‘‡ ๐‘šโˆ’1 ) 2 = ๐ผ + 2๐‘Ž 1 ๐‘‡ + (2๐‘Ž 2 + ๐‘Ž 12 )๐‘‡ 2 + (2๐‘Ž 3 + 2๐‘Ž 1 ๐‘Ž 2 )๐‘‡ 3 + โ‹ฏ + (2๐‘Ž ๐‘šโˆ’1 + terms involving ๐‘Ž 1 , โ€ฆ , ๐‘Ž ๐‘šโˆ’2 )๐‘‡ ๐‘šโˆ’1 . We want the right side of the equation above to equal ๐ผ + ๐‘‡ . Hence choose ๐‘Ž 1 such that 2๐‘Ž 1 = 1 (thus ๐‘Ž 1 = 1/2 ). Next, choose ๐‘Ž 2 such that 2๐‘Ž 2 + ๐‘Ž 12 = 0 (thus ๐‘Ž 2 = โˆ’1/8 ). Then choose ๐‘Ž 3 such that the coefficient of ๐‘‡ 3 on the right side of the equation above equals 0 (thus ๐‘Ž 3 = 1/16 ). Continue in this fashion for each ๐‘˜ = 4 , โ€ฆ , ๐‘šโˆ’1 , at each step solving for ๐‘Ž ๐‘˜ so that the coefficient of ๐‘‡ ๐‘˜ on the right side of the equation above equals 0 . Actually we do not care about the explicit formula for the ๐‘Ž ๐‘˜ โ€™s. We only need to know that some choice of the ๐‘Ž ๐‘˜ โ€™s gives a square root of ๐ผ + ๐‘‡ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 333 Spans: True Boxes: True Text: 320 Chapter 8 Operators on Complex Vector Spaces The previous lemma is valid on real and complex vector spaces. However, the result below holds only on complex vector spaces. For example, the operator of multiplication by โˆ’1 on the one-dimensional real vector space ๐‘ has no square root. Representation of a complex number with polar coordinates. For the proof below, we need to know that every ๐‘ง โˆˆ ๐‚ has a square root in ๐‚ . To show this, write ๐‘ง = ๐‘Ÿ( cos ๐œƒ + ๐‘– sin ๐œƒ) , where ๐‘Ÿ is the length of the line segment in the complex plane from the origin to ๐‘ง and ๐œƒ is the angle of that line segment with the positive horizontal axis. Then โˆš ๐‘Ÿ( cos ๐œƒ2 + ๐‘– sin ๐œƒ2 ) is a square root of ๐‘ง , as you can verify by showing that the square of the complex number above equals ๐‘ง . 8.41 over ๐‚ , invertible operators have square roots Suppose ๐‘‰ is a complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible. Then ๐‘‡ has a square root. Proof Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š be the distinct eigenvalues of ๐‘‡ . For each ๐‘˜ , there exists a nilpotent operator ๐‘‡ ๐‘˜ โˆˆ โ„’ (๐บ(๐œ† ๐‘˜ , ๐‘‡)) such that ๐‘‡| ๐บ(๐œ† ๐‘˜ , ๐‘‡) = ๐œ† ๐‘˜ ๐ผ + ๐‘‡ ๐‘˜ [ see 8.22(b) ] . Because ๐‘‡ is invertible, none of the ๐œ† ๐‘˜ โ€™s equals 0 , so we can write ๐‘‡| ๐บ(๐œ† ๐‘˜ , ๐‘‡) = ๐œ† ๐‘˜ (๐ผ + ๐‘‡ ๐‘˜ ๐œ† ๐‘˜ ) for each ๐‘˜ . Because ๐‘‡ ๐‘˜ /๐œ† ๐‘˜ is nilpotent, ๐ผ + ๐‘‡ ๐‘˜ /๐œ† ๐‘˜ has a square root (by 8.39). Multiplying a square root of the complex number ๐œ† ๐‘˜ by a square root of ๐ผ + ๐‘‡ ๐‘˜ /๐œ† ๐‘˜ , we obtain a square root ๐‘… ๐‘˜ of ๐‘‡| ๐บ(๐œ† ๐‘˜ , ๐‘‡) . By the generalized eigenspace decomposition (8.22), a typical vector ๐‘ฃ โˆˆ ๐‘‰ can be written uniquely in the form ๐‘ฃ = ๐‘ข 1 + โ‹ฏ + ๐‘ข ๐‘š , where each ๐‘ข ๐‘˜ is in ๐บ(๐œ† ๐‘˜ , ๐‘‡) . Using this decomposition, define an operator ๐‘… โˆˆ โ„’ (๐‘‰) by ๐‘…๐‘ฃ = ๐‘… 1 ๐‘ข 1 + โ‹ฏ + ๐‘… ๐‘š ๐‘ข ๐‘š . You should verify that this operator ๐‘… is a square root of ๐‘‡ , completing the proof. By imitating the techniques in this subsection, you should be able to prove that if ๐‘‰ is a complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible, then ๐‘‡ has a ๐‘˜ th root for every positive integer ๐‘˜ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 334 Spans: True Boxes: True Text: Section 8C Consequences of Generalized Eigenspace Decomposition 321 Jordan Form We know that if ๐‘‰ is a complex vector space, then for every ๐‘‡ โˆˆ โ„’ (๐‘‰) there is a basis of ๐‘‰ with respect to which ๐‘‡ has a nice upper-triangular matrix (see 8.37). In this subsection we will see that we can do even betterโ€”there is a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ contains 0 โ€™s everywhere except possibly on the diagonal and the line directly above the diagonal. We begin by looking at two examples of nilpotent operators. 8.42 example: nilpotent operator with nice matrix Let ๐‘‡ be the operator on ๐‚ 4 defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 ) = (0 , ๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ). Then ๐‘‡ 4 = 0 ; thus ๐‘‡ is nilpotent. If ๐‘ฃ = (1 , 0 , 0 , 0) , then ๐‘‡ 3 ๐‘ฃ , ๐‘‡ 2 ๐‘ฃ , ๐‘‡๐‘ฃ , ๐‘ฃ is a basis of ๐‚ 4 . The matrix of ๐‘‡ with respect to this basis is โŽ›โŽœโŽœโŽœโŽœโŽœโŽœโŽ 0 1 0 0 0 0 1 0 0 0 0 1 0 0 0 0 โŽžโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽ . The next example of a nilpotent operator has more complicated behavior than the example above. 8.43 example: nilpotent operator with slightly more complicated matrix Let ๐‘‡ be the operator on ๐‚ 6 defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 , ๐‘ง 5 , ๐‘ง 6 ) = (0 , ๐‘ง 1 , ๐‘ง 2 , 0 , ๐‘ง 4 , 0). Then ๐‘‡ 3 = 0 ; thus ๐‘‡ is nilpotent. In contrast to the nice behavior of the nilpotent operator of the previous example, for this nilpotent operator there does not exist a vector ๐‘ฃ โˆˆ ๐‚ 6 such that ๐‘‡ 5 ๐‘ฃ , ๐‘‡ 4 ๐‘ฃ , ๐‘‡ 3 ๐‘ฃ , ๐‘‡ 2 ๐‘ฃ , ๐‘‡๐‘ฃ , ๐‘ฃ is a basis of ๐‚ 6 . However, if we take ๐‘ฃ 1 = (1 , 0 , 0 , 0 , 0 , 0) , ๐‘ฃ 2 = (0 , 0 , 0 , 1 , 0 , 0) , and ๐‘ฃ 3 = (0 , 0 , 0 , 0 , 0 , 1) , then ๐‘‡ 2 ๐‘ฃ 1 , ๐‘‡๐‘ฃ 1 , ๐‘ฃ 1 , ๐‘‡๐‘ฃ 2 , ๐‘ฃ 2 , ๐‘ฃ 3 is a basis of ๐‚ 6 . The matrix of ๐‘‡ with respect to this basis is โŽ›โŽœโŽœโŽœโŽœโŽœ โŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽ โŽ›โŽœโŽœโŽœ โŽ 0 1 0 0 0 1 0 0 0 โŽžโŽŸโŽŸโŽŸ โŽ  0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ( 0 1 0 0 ) 0 0 0 0 0 0 0 ( 0 ) โŽžโŽŸโŽŸโŽŸโŽŸโŽŸ โŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽ  . Here the inner matrices are blocked off to show that we can think of the 6 -by- 6 matrix above as a block diagonal matrix consisting of a 3 -by- 3 block with 1 โ€™s on the line above the diagonal and 0 โ€™s elsewhere, a 2 -by- 2 block with 1 above the diagonal and 0 โ€™s elsewhere, and a 1 -by- 1 block containing 0 . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 335 Spans: True Boxes: True Text: 322 Chapter 8 Operators on Complex Vector Spaces Our next goal is to show that every nilpotent operator ๐‘‡ โˆˆ โ„’ (๐‘‰) behaves similarly to the operator in the previous example. Specifically, there is a finite collection of vectors ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› โˆˆ ๐‘‰ such that there is a basis of ๐‘‰ consisting of the vectors of the form ๐‘‡ ๐‘— ๐‘ฃ ๐‘˜ , as ๐‘˜ varies from 1 to ๐‘› and ๐‘— varies (in reverse order) from 0 to the largest nonnegative integer ๐‘š ๐‘˜ such that ๐‘‡ ๐‘š ๐‘˜ ๐‘ฃ ๐‘˜ โ‰  0 . With respect to this basis, the matrix of ๐‘‡ looks like the matrix in the previous example. More specifically, ๐‘‡ has a block diagonal matrix with respect to this basis, with each block a square matrix that is 0 everywhere except on the line above the diagonal. In the next definition, the diagonal of each ๐ด ๐‘˜ is filled with some eigenvalue ๐œ† ๐‘˜ of ๐‘‡ , the line directly above the diagonal of ๐ด ๐‘˜ is filled with 1 โ€™s, and all other entries in ๐ด ๐‘˜ are 0 (to understand why each ๐œ† ๐‘˜ is an eigenvalue of ๐‘‡ , see 5.41). The ๐œ† ๐‘˜ โ€™s need not be distinct. Also, ๐ด ๐‘˜ may be a 1 -by- 1 matrix (๐œ† ๐‘˜ ) containing just an eigenvalue of ๐‘‡ . If each ๐œ† ๐‘˜ is 0 , then the next definition captures the behavior described in the paragraph above (recall that if ๐‘‡ is nilpotent, then 0 is the only eigenvalue of ๐‘‡ ). 8.44 definition: Jordan basis Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . A basis of ๐‘‰ is called a Jordan basis for ๐‘‡ if with respect to this basis ๐‘‡ has a block diagonal matrix โŽ›โŽœโŽœโŽœโŽ ๐ด 1 0 โ‹ฑ 0 ๐ด ๐‘ โŽžโŽŸโŽŸโŽŸโŽ  in which each ๐ด ๐‘˜ is an upper-triangular matrix of the form ๐ด ๐‘˜ = โŽ› โŽœ โŽœโŽœโŽœโŽœโŽœโŽ ๐œ† ๐‘˜ 1 0 โ‹ฑ โ‹ฑ โ‹ฑ 1 0 ๐œ† ๐‘˜ โŽž โŽŸ โŽŸโŽŸโŽŸโŽŸโŽŸโŽ . Most of the work in proving that every operator on a finite-dimensional com- plex vector space has a Jordan basis occurs in proving the special case below of nilpotent operators. This special case holds on real vector spaces as well as complex vector spaces. 8.45 every nilpotent operator has a Jordan basis Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Then there is a basis of ๐‘‰ that is a Jordan basis for ๐‘‡ . Proof We will prove this result by induction on dim ๐‘‰ . To get started, note that the desired result holds if dim ๐‘‰ = 1 (because in that case, the only nilpotent operator is the 0 operator). Now assume that dim ๐‘‰ > 1 and that the desired result holds on all vector spaces of smaller dimension. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 336 Spans: True Boxes: True Text: Section 8C Consequences of Generalized Eigenspace Decomposition 323 Let ๐‘š be the smallest positive integer such that ๐‘‡ ๐‘š = 0 . Thus there exists ๐‘ข โˆˆ ๐‘‰ such that ๐‘‡ ๐‘šโˆ’1 ๐‘ข โ‰  0 . Let ๐‘ˆ = span (๐‘ข , ๐‘‡๐‘ข , โ€ฆ , ๐‘‡ ๐‘šโˆ’1 ๐‘ข). The list ๐‘ข , ๐‘‡๐‘ข , โ€ฆ , ๐‘‡ ๐‘šโˆ’1 ๐‘ข is linearly independent (see Exercise 2 in Section 8A). If ๐‘ˆ = ๐‘‰ , then writing this list in reverse order gives a Jordan basis for ๐‘‡ and we are done. Thus we can assume that ๐‘ˆ โ‰  ๐‘‰ . Note that ๐‘ˆ is invariant under ๐‘‡ . By our induction hypothesis, there is a basis of ๐‘ˆ that is a Jordan basis for ๐‘‡| ๐‘ˆ . The strategy of our proof is that we will find a subspace ๐‘Š of ๐‘‰ such that ๐‘Š is also invariant under ๐‘‡ and ๐‘‰ = ๐‘ˆ โŠ• ๐‘Š . Again by our induction hypothesis, there will be a basis of ๐‘Š that is a Jordan basis for ๐‘‡| ๐‘Š . Putting together the Jordan bases for ๐‘‡| ๐‘ˆ and ๐‘‡| ๐‘Š , we will have a Jordan basis for ๐‘‡ . Let ๐œ‘ โˆˆ ๐‘‰ โ€ฒ be such that ๐œ‘(๐‘‡ ๐‘šโˆ’1 ๐‘ข) โ‰  0 . Let ๐‘Š = {๐‘ฃ โˆˆ ๐‘‰ โˆถ ๐œ‘(๐‘‡ ๐‘˜ ๐‘ฃ) = 0 for each ๐‘˜ = 0 , โ€ฆ , ๐‘š โˆ’ 1}. Then ๐‘Š is a subspace of ๐‘‰ that is invariant under ๐‘‡ ( the invariance holds because if ๐‘ฃ โˆˆ ๐‘Š then ๐œ‘(๐‘‡ ๐‘˜ (๐‘‡๐‘ฃ)) = 0 for ๐‘˜ = 0 , โ€ฆ , ๐‘š โˆ’ 1 , where the case ๐‘˜ = ๐‘š โˆ’ 1 holds because ๐‘‡ ๐‘š = 0) . We will show that ๐‘‰ = ๐‘ˆ โŠ• ๐‘Š , which by the previous paragraph will complete the proof. To show that ๐‘ˆ + ๐‘Š is a direct sum, suppose ๐‘ฃ โˆˆ ๐‘ˆ โˆฉ ๐‘Š with ๐‘ฃ โ‰  0 . Because ๐‘ฃ โˆˆ ๐‘ˆ , there exist ๐‘ 0 , โ€ฆ , ๐‘ ๐‘šโˆ’1 โˆˆ ๐… such that ๐‘ฃ = ๐‘ 0 ๐‘ข + ๐‘ 1 ๐‘‡๐‘ข + โ‹ฏ + ๐‘ ๐‘šโˆ’1 ๐‘‡ ๐‘šโˆ’1 ๐‘ข. Let ๐‘— be the smallest index such that ๐‘ ๐‘— โ‰  0 . Apply ๐‘‡ ๐‘šโˆ’๐‘—โˆ’1 to both sides of the equation above, getting ๐‘‡ ๐‘šโˆ’๐‘—โˆ’1 ๐‘ฃ = ๐‘ ๐‘— ๐‘‡ ๐‘šโˆ’1 ๐‘ข , where we have used the equation ๐‘‡ ๐‘š = 0 . Now apply ๐œ‘ to both sides of the equation above, getting ๐œ‘(๐‘‡ ๐‘šโˆ’๐‘—โˆ’1 ๐‘ฃ) = ๐‘ ๐‘— ๐œ‘(๐‘‡ ๐‘šโˆ’1 ๐‘ข) โ‰  0. The equation above shows that ๐‘ฃ โˆ‰ ๐‘Š . Hence we have proved that ๐‘ˆ โˆฉ ๐‘Š = {0} , which implies that ๐‘ˆ + ๐‘Š is a direct sum (see 1.46). To show that ๐‘ˆ โŠ• ๐‘Š = ๐‘‰ , define ๐‘† โˆถ ๐‘‰ โ†’ ๐… ๐‘š by ๐‘†๐‘ฃ = (๐œ‘(๐‘ฃ) , ๐œ‘(๐‘‡๐‘ฃ) , โ€ฆ , ๐œ‘(๐‘‡ ๐‘šโˆ’1 ๐‘ฃ)). Thus null ๐‘† = ๐‘Š . Hence dim ๐‘Š = dim null ๐‘† = dim ๐‘‰ โˆ’ dim range ๐‘† โ‰ฅ dim ๐‘‰ โˆ’ ๐‘š , where the second equality comes from the fundamental theorem of linear maps (3.21). Using the inequality above, we have dim (๐‘ˆ โŠ• ๐‘Š) = dim ๐‘ˆ + dim ๐‘Š โ‰ฅ ๐‘š + ( dim ๐‘‰ โˆ’ ๐‘š) = dim ๐‘‰. Thus ๐‘ˆ โŠ• ๐‘Š = ๐‘‰ (by 2.39), completing the proof. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 337 Spans: True Boxes: True Text: 324 Chapter 8 Operators on Complex Vector Spaces Camille Jordan ( 1838โ€“1922 ) pub- lished a proof of 8.46 in 1870. Now the generalized eigenspace de- composition allows us to extend the pre- vious result to operators that may not be nilpotent. Doing this requires that we deal with complex vector spaces. 8.46 Jordan form Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then there is a basis of ๐‘‰ that is a Jordan basis for ๐‘‡ . Proof Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š be the distinct eigenvalues of ๐‘‡ . The generalized eigenspace decomposition states that ๐‘‰ = ๐บ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐บ(๐œ† ๐‘š , ๐‘‡) , where each (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) is nilpotent (see 8.22). Thus 8.45 implies that some basis of each ๐บ(๐œ† ๐‘˜ , ๐‘‡) is a Jordan basis for (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)| ๐บ(๐œ† ๐‘˜ , ๐‘‡) . Put these bases together to get a basis of ๐‘‰ that is a Jordan basis for ๐‘‡ . Exercises 8C 1 Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) is the operator defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (๐‘ง 2 , ๐‘ง 3 , 0) . Prove that ๐‘‡ does not have a square root. 2 Define ๐‘‡ โˆˆ โ„’ (๐… 5 ) by ๐‘‡(๐‘ฅ 1 , ๐‘ฅ 2 , ๐‘ฅ 3 , ๐‘ฅ 4 , ๐‘ฅ 5 ) = (2๐‘ฅ 2 , 3๐‘ฅ 3 , โˆ’๐‘ฅ 4 , 4๐‘ฅ 5 , 0) . (a) Show that ๐‘‡ is nilpotent. (b) Find a square root of ๐ผ + ๐‘‡ . 3 Suppose ๐‘‰ is a complex vector space. Prove that every invertible operator on ๐‘‰ has a cube root. 4 Suppose ๐‘‰ is a real vector space. Prove that the operator โˆ’๐ผ on ๐‘‰ has a square root if and only if dim ๐‘‰ is an even number. 5 Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 2 ) is the operator defined by ๐‘‡(๐‘ค , ๐‘ง) = (โˆ’๐‘คโˆ’๐‘ง , 9๐‘ค + 5๐‘ง) . Find a Jordan basis for ๐‘‡ . 6 Find a basis of ๐’ซ 4 (๐‘) that is a Jordan basis for the differentiation operator ๐ท on ๐’ซ 4 (๐‘) defined by ๐ท๐‘ = ๐‘ โ€ฒ . 7 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a Jordan basis for ๐‘‡ . Prove that the minimal polynomial of ๐‘‡ is ๐‘ง ๐‘š + 1 , where ๐‘š is the length of the longest consecutive string of 1 โ€™s that appears on the line directly above the diagonal in the matrix of ๐‘‡ with respect to ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› . 8 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ that is a Jordan basis for ๐‘‡ . Describe the matrix of ๐‘‡ 2 with respect to this basis. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 338 Spans: True Boxes: True Text: Section 8C Consequences of Generalized Eigenspace Decomposition 325 9 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Explain why there exist ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› โˆˆ ๐‘‰ and nonnegative integers ๐‘š 1 , โ€ฆ , ๐‘š ๐‘› such that (a) and (b) below both hold. (a) ๐‘‡ ๐‘š 1 ๐‘ฃ 1 , โ€ฆ , ๐‘‡๐‘ฃ 1 , ๐‘ฃ 1 , โ€ฆ , ๐‘‡ ๐‘š ๐‘› ๐‘ฃ ๐‘› , โ€ฆ , ๐‘‡๐‘ฃ ๐‘› , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . (b) ๐‘‡ ๐‘š 1 + 1 ๐‘ฃ 1 = โ‹ฏ = ๐‘‡ ๐‘š ๐‘› + 1 ๐‘ฃ ๐‘› = 0 . 10 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ that is a Jordan basis for ๐‘‡ . Describe the matrix of ๐‘‡ with respect to the basis ๐‘ฃ ๐‘› , โ€ฆ , ๐‘ฃ 1 obtained by reversing the order of the ๐‘ฃ โ€™s. 11 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Explain why every vector in each Jordan basis for ๐‘‡ is a generalized eigenvector of ๐‘‡ . 12 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is diagonalizable. Show that โ„ณ (๐‘‡) is a diagonal matrix with respect to every Jordan basis for ๐‘‡ . 13 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is nilpotent. Prove that if ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› are vectors in ๐‘‰ and ๐‘š 1 , โ€ฆ , ๐‘š ๐‘› are nonnegative integers such that ๐‘‡ ๐‘š 1 ๐‘ฃ 1 , โ€ฆ , ๐‘‡๐‘ฃ 1 , ๐‘ฃ 1 , โ€ฆ , ๐‘‡ ๐‘š ๐‘› ๐‘ฃ ๐‘› , โ€ฆ , ๐‘‡๐‘ฃ ๐‘› , ๐‘ฃ ๐‘› is a basis of ๐‘‰ and ๐‘‡ ๐‘š 1 + 1 ๐‘ฃ 1 = โ‹ฏ = ๐‘‡ ๐‘š ๐‘› + 1 ๐‘ฃ ๐‘› = 0 , then ๐‘‡ ๐‘š 1 ๐‘ฃ 1 , โ€ฆ , ๐‘‡ ๐‘š ๐‘› ๐‘ฃ ๐‘› is a basis of null ๐‘‡ . This exercise shows that ๐‘› = dim null ๐‘‡ . Thus the positive integer ๐‘› that appears above depends only on ๐‘‡ and not on the specific Jordan basis chosen for ๐‘‡ . 14 Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that there does not exist a direct sum decomposition of ๐‘‰ into two nonzero subspaces invariant under ๐‘‡ if and only if the minimal polynomial of ๐‘‡ is of the form (๐‘ง โˆ’ ๐œ†) dim ๐‘‰ for some ๐œ† โˆˆ ๐‚ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 339 Spans: True Boxes: True Text: 326 Chapter 8 Operators on Complex Vector Spaces 8D Trace: A Connection Between Matrices and Operators We begin this section by defining the trace of a square matrix. After developing some properties of the trace of a square matrix, we will use this concept to define the trace of an operator. 8.47 definition: trace of a matrix Suppose ๐ด is a square matrix with entries in ๐… . The trace of ๐ด , denoted by tr ๐ด , is defined to be the sum of the diagonal entries of ๐ด . 8.48 example: trace of a 3 -by- 3 matrix Suppose ๐ด = โŽ›โŽœโŽœโŽœโŽ 3 โˆ’1 โˆ’2 3 2 โˆ’3 1 2 0 โŽžโŽŸโŽŸโŽŸโŽ . The diagonal entries of ๐ด , which are shown in red above, are 3 , 2 , and 0 . Thus tr ๐ด = 3 + 2 + 0 = 5 . Matrix multiplication is not commutative, but the next result shows that the order of matrix multiplication does not matter to the trace. 8.49 trace of ๐ด๐ต equals trace of ๐ต๐ด Suppose ๐ด is an ๐‘š -by- ๐‘› matrix and ๐ต is an ๐‘› -by- ๐‘š matrix. Then tr (๐ด๐ต) = tr (๐ต๐ด). Proof Suppose ๐ด = โŽ›โŽœโŽœโŽœโŽœโŽ ๐ด 1 , 1 โ‹ฏ ๐ด 1 , ๐‘› โ‹ฎ โ‹ฎ ๐ด ๐‘š , 1 โ‹ฏ ๐ด ๐‘š , ๐‘› โŽžโŽŸโŽŸโŽŸโŽŸโŽ  , ๐ต = โŽ›โŽœโŽœโŽœโŽœโŽ ๐ต 1 , 1 โ‹ฏ ๐ต 1 , ๐‘š โ‹ฎ โ‹ฎ ๐ต ๐‘› , 1 โ‹ฏ ๐ต ๐‘› , ๐‘š โŽžโŽŸโŽŸโŽŸโŽŸโŽ . The ๐‘— th term on the diagonal of the ๐‘š -by- ๐‘š matrix ๐ด๐ต equals โˆ‘ ๐‘›๐‘˜=1 ๐ด ๐‘— , ๐‘˜ ๐ต ๐‘˜ , ๐‘— . Thus tr (๐ด๐ต) = ๐‘š โˆ‘ ๐‘— = 1 ๐‘› โˆ‘ ๐‘˜=1 ๐ด ๐‘— , ๐‘˜ ๐ต ๐‘˜ , ๐‘— = ๐‘› โˆ‘ ๐‘˜=1 ๐‘š โˆ‘ ๐‘— = 1 ๐ต ๐‘˜ , ๐‘— ๐ด ๐‘— , ๐‘˜ = ๐‘› โˆ‘ ๐‘˜=1 (๐‘˜ th term on diagonal of the ๐‘› -by- ๐‘› matrix ๐ต๐ด) = tr (๐ต๐ด) , as desired. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 340 Spans: True Boxes: True Text: Section 8D Trace: A Connection Between Matrices and Operators 327 We want to define the trace of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) to be the trace of the matrix of ๐‘‡ with respect to some basis of ๐‘‰ . However, this definition should not depend on the choice of basis. The following result will make this possible. 8.50 trace of matrix of operator does not depend on basis Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Suppose ๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘› and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› are bases of ๐‘‰ . Then tr โ„ณ (๐‘‡ , (๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘› )) = tr โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )). Proof Let ๐ด = โ„ณ (๐‘‡ , (๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘› )) and ๐ต = โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) . The change-of- basis formula tells us that there exists an invertible ๐‘› -by- ๐‘› matrix ๐ถ such that ๐ด = ๐ถ โˆ’1 ๐ต๐ถ (see 3.84). Thus tr ๐ด = tr ((๐ถ โˆ’1 ๐ต)๐ถ) = tr (๐ถ(๐ถ โˆ’1 ๐ต)) = tr ((๐ถ๐ถ โˆ’1 )๐ต) = tr ๐ต , where the second line comes from 8.49. Because of 8.50, the following definition now makes sense. 8.51 definition: trace of an operator Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . The trace of ๐‘‡ , denote tr ๐‘‡ , is defined by tr ๐‘‡ = tr โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) , where ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is any basis of ๐‘‰ . Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† is an eigenvalue of ๐‘‡ . Recall that we defined the multiplicity of ๐œ† to be the dimension of the generalized eigenspace ๐บ(๐œ† , ๐‘‡) (see 8.23); we proved that this multiplicity equals dim null (๐‘‡ โˆ’ ๐œ†๐ผ) dim ๐‘‰ (see 8.20). Recall also that if ๐‘‰ is a complex vector space, then the sum of the multiplicities of all eigenvalues of ๐‘‡ equals dim ๐‘‰ (see 8.25). In the definition below, the sum of the eigenvalues โ€œwith each eigenvalue included as many times as its multiplicityโ€ means that if ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are the distinct eigenvalues of ๐‘‡ with multiplicities ๐‘‘ 1 , โ€ฆ , ๐‘‘ ๐‘š , then the sum is ๐‘‘ 1 ๐œ† 1 + โ‹ฏ + ๐‘‘ ๐‘š ๐œ† ๐‘š . Or if you prefer to work with a list of not-necessarily-distinct eigenvalues, with each eigenvalue included as many times as its multiplicity, then the eigenvalues could be denoted by ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› (where ๐‘› equals dim ๐‘‰ ) and the sum is ๐œ† 1 + โ‹ฏ + ๐œ† ๐‘› . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 341 Spans: True Boxes: True Text: 328 Chapter 8 Operators on Complex Vector Spaces 8.52 on complex vector spaces, trace equals sum of eigenvalues Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then tr ๐‘‡ equals the sum of the eigenvalues of ๐‘‡ , with each eigenvalue included as many times as its multiplicity. Proof There is a basis of ๐‘‰ with respect to which ๐‘‡ has an upper-triangular matrix with the diagonal entries of the matrix consisting of the eigenvalues of ๐‘‡ , with each eigenvalue included as many times as its multiplicityโ€”see 8.37. Thus the definition of the trace of an operator along with 8.50, which allows us to use a basis of our choice, implies that tr ๐‘‡ equals the sum of the eigenvalues of ๐‘‡ , with each eigenvalue included as many times as its multiplicity. 8.53 example: trace of an operator on ๐‚ 3 Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) is defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (3๐‘ง 1 โˆ’ ๐‘ง 2 โˆ’ 2๐‘ง 3 , 3๐‘ง 1 + 2๐‘ง 2 โˆ’ 3๐‘ง 3 , ๐‘ง 1 + 2๐‘ง 2 ). Then the matrix of ๐‘‡ with respect to the standard basis of ๐‚ 3 is โŽ›โŽœโŽœโŽœโŽ 3 โˆ’1 โˆ’2 3 2 โˆ’3 1 2 0 โŽžโŽŸโŽŸโŽŸโŽ . Adding up the diagonal entries of this matrix, we see that tr ๐‘‡ = 5 . The eigenvalues of ๐‘‡ are 1 , 2 + 3๐‘– , and 2 โˆ’ 3๐‘– , each with multiplicity 1 , as you can verify. The sum of these eigenvalues, each included as many times as its multiplicity, is 1 + (2 + 3๐‘–) + (2 โˆ’ 3๐‘–) , which equals 5 , as expected by 8.52. The trace has a close connection with the characteristic polynomial. Suppose ๐… = ๐‚ , ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› are the eigenvalues of ๐‘‡ , with each eigenvalue included as many times as its multiplicity. Then by definition (see 8.26), the characteristic polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘› ). Expanding the polynomial above, we can write the characteristic polynomial of ๐‘‡ in the form ๐‘ง ๐‘› โˆ’ (๐œ† 1 + โ‹ฏ + ๐œ† ๐‘› )๐‘ง ๐‘›โˆ’1 + โ‹ฏ + (โˆ’1) ๐‘› (๐œ† 1 โ‹ฏ๐œ† ๐‘› ). The expression above immediately leads to the next result. Also see 9.65, which does not require the hypothesis that ๐… = ๐‚ . 8.54 trace and characteristic polynomial Suppose ๐… = ๐‚ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐‘› = dim ๐‘‰ . Then tr ๐‘‡ equals the negative of the coefficient of ๐‘ง ๐‘›โˆ’1 in the characteristic polynomial of ๐‘‡ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 342 Spans: True Boxes: True Text: Section 8D Trace: A Connection Between Matrices and Operators 329 The next result gives a nice formula for the trace of an operator on an inner product space. 8.55 trace on an inner product space Suppose ๐‘‰ is an inner product space, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘’ 1 , โ€ฆ , ๐‘’ ๐‘› is an orthonor- mal basis of ๐‘‰ . Then tr ๐‘‡ = โŸจ๐‘‡๐‘’ 1 , ๐‘’ 1 โŸฉ + โ‹ฏ + โŸจ๐‘‡๐‘’ ๐‘› , ๐‘’ ๐‘› โŸฉ. Proof The desired formula follows from the observation that the entry in row ๐‘˜ , column ๐‘˜ of โ„ณ (๐‘‡ , (๐‘’ 1 , โ€ฆ , ๐‘’ ๐‘› )) equals โŸจ๐‘‡๐‘’ ๐‘˜ , ๐‘’ ๐‘˜ โŸฉ [use 6.30(a) with ๐‘ฃ = ๐‘‡๐‘’ ๐‘˜ ]. The algebraic properties of the trace as defined on square matrices translate to algebraic properties of the trace as defined on operators, as shown in the next result. 8.56 trace is linear The function tr โˆถ โ„’ (๐‘‰) โ†’ ๐… is a linear functional on โ„’ (๐‘‰) such that tr (๐‘†๐‘‡) = tr (๐‘‡๐‘†) for all ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) . Proof Choose a basis of ๐‘‰ . All matrices of operators in this proof will be with respect to that basis. Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) . If ๐œ† โˆˆ ๐… , then tr (๐œ†๐‘‡) = tr โ„ณ (๐œ†๐‘‡) = tr (๐œ† โ„ณ (๐‘‡)) = ๐œ† tr โ„ณ (๐‘‡) = ๐œ† tr ๐‘‡ , where the first and last equalities come from the definition of the trace of an operator, the second equality comes from 3.38, and the third equality follows from the definition of the trace of a square matrix. Also, tr (๐‘† + ๐‘‡) = tr โ„ณ (๐‘† + ๐‘‡) = tr ( โ„ณ (๐‘†) + โ„ณ (๐‘‡)) = tr โ„ณ (๐‘†) + tr โ„ณ (๐‘‡) = tr ๐‘† + tr ๐‘‡ , where the first and last equalities come from the definition of the trace of an operator, the second equality comes from 3.35, and the third equality follows from the definition of the trace of a square matrix. The two paragraphs above show that tr โˆถ โ„’ (๐‘‰) โ†’ ๐… is a linear functional on โ„’ (๐‘‰) . Furthermore, tr (๐‘†๐‘‡) = tr โ„ณ (๐‘†๐‘‡) = tr ( โ„ณ (๐‘†) โ„ณ (๐‘‡)) = tr ( โ„ณ (๐‘‡) โ„ณ (๐‘†)) = tr โ„ณ (๐‘‡๐‘†) = tr (๐‘‡๐‘†) , where the second and fourth equalities come from 3.43 and the crucial third equality comes from 8.49. The equations tr (๐‘†๐‘‡) = tr (๐‘‡๐‘†) and tr ๐ผ = dim ๐‘‰ uniquely characterize the trace among the linear functionals on โ„’ (๐‘‰) โ€”see Exercise 10. Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 343 Spans: True Boxes: True Text: 330 Chapter 8 Operators on Complex Vector Spaces The statement of the next result does not involve traces, but the short proof uses traces. When something like this happens in mathematics, then usually a good definition lurks in the back- ground. The equation tr (๐‘†๐‘‡) = tr (๐‘‡๐‘†) leads to our next result, which does not hold on infinite-dimensional vector spaces (see Exercise 13). However, additional hy- potheses on ๐‘† , ๐‘‡ , and ๐‘‰ lead to an infinite- dimensional generalization of the result below, with important applications to quantum theory. 8.57 identity operator is not the difference of ๐‘†๐‘‡ and ๐‘‡๐‘† There do not exist operators ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) such that ๐‘†๐‘‡ โˆ’ ๐‘‡๐‘† = ๐ผ . Proof Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then tr (๐‘†๐‘‡ โˆ’ ๐‘‡๐‘†) = tr (๐‘†๐‘‡) โˆ’ tr (๐‘‡๐‘†) = 0 , where both equalities come from 8.56. The trace of ๐ผ equals dim ๐‘‰ , which is not 0 . Because ๐‘†๐‘‡ โˆ’ ๐‘‡๐‘† and ๐ผ have different traces, they cannot be equal. Exercises 8D 1 Suppose ๐‘‰ is an inner product space and ๐‘ฃ , ๐‘ค โˆˆ ๐‘‰ . Define an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) by ๐‘‡๐‘ข = โŸจ๐‘ข , ๐‘ฃโŸฉ๐‘ค . Find a formula for tr ๐‘‡ . 2 Suppose ๐‘ƒ โˆˆ โ„’ (๐‘‰) satisfies ๐‘ƒ 2 = ๐‘ƒ . Prove that tr ๐‘ƒ = dim range ๐‘ƒ. 3 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘‡ 5 = ๐‘‡ . Prove that the real and imaginary parts of tr ๐‘‡ are both integers. 4 Suppose ๐‘‰ is an inner product space and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that tr ๐‘‡ โˆ— = tr ๐‘‡. 5 Suppose ๐‘‰ is an inner product space. Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is a positive operator and tr ๐‘‡ = 0 . Prove that ๐‘‡ = 0 . 6 Suppose ๐‘‰ is an inner product space and ๐‘ƒ , ๐‘„ โˆˆ โ„’ (๐‘‰) are orthogonal projections. Prove that tr (๐‘ƒ๐‘„) โ‰ฅ 0 . 7 Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) is the operator whose matrix is โŽ›โŽœโŽœโŽœโŽ 51 โˆ’12 โˆ’21 60 โˆ’40 โˆ’28 57 โˆ’68 1 โŽžโŽŸโŽŸโŽŸโŽ . Someone tells you (accurately) that โˆ’48 and 24 are eigenvalues of ๐‘‡ . Without using a computer or writing anything down, find the third eigenvalue of ๐‘‡ . Linear Algebra Done Right , fourth edition, by Sheldon Axler Annotated Entity: ID: 344 Spans: True Boxes: True Text: Section 8D Trace: A Connection Between Matrices and Operators 331 8 Prove or give a counterexample: If ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) , then tr (๐‘†๐‘‡) = ( tr ๐‘†)( tr ๐‘‡) . 9 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that tr (๐‘†๐‘‡) = 0 for all ๐‘† โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ = 0 . 10 Prove that the trace is the only linear functional ๐œ โˆถ โ„’ (๐‘‰) โ†’ ๐… such that ๐œ(๐‘†๐‘‡) = ๐œ(๐‘‡๐‘†) for all ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ(๐ผ) = dim ๐‘‰ . Hint: Suppose that ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . For ๐‘— , ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘›} , define ๐‘ƒ ๐‘— , ๐‘˜ โˆˆ โ„’ (๐‘‰) by ๐‘ƒ ๐‘— , ๐‘˜ (๐‘Ž 1 ๐‘ฃ 1 + โ‹ฏ + ๐‘Ž ๐‘› ๐‘ฃ ๐‘› ) = ๐‘Ž ๐‘˜ ๐‘ฃ ๐‘— . Prove that ๐œ(๐‘ƒ ๐‘— , ๐‘˜ ) = โŽง{ โŽจ{โŽฉ 1 if ๐‘— = ๐‘˜ , 0 if ๐‘— โ‰  ๐‘˜. Then for ๐‘‡ โˆˆ โ„’ (๐‘‰) , use the equation ๐‘‡ = โˆ‘ ๐‘›๐‘˜=1 โˆ‘ ๐‘› ๐‘— = 1 โ„ณ (๐‘‡) ๐‘— , ๐‘˜ ๐‘ƒ ๐‘— , ๐‘˜ to show that ๐œ(๐‘‡) = tr ๐‘‡ . 11 Suppose ๐‘‰ and ๐‘Š are inner product spaces and ๐‘‡ โˆˆ โ„’ (๐‘‰ , ๐‘Š) . Prove that if ๐‘’ 1 , โ€ฆ , ๐‘’ ๐‘› is an orthonormal basis of ๐‘‰ and ๐‘“ 1 , โ€ฆ , ๐‘“ ๐‘š is an orthonormal basis of ๐‘Š , then tr (๐‘‡ โˆ— ๐‘‡) = ๐‘› โˆ‘ ๐‘˜=1 ๐‘š โˆ‘ ๐‘— = 1 |โŸจ๐‘‡๐‘’ ๐‘˜ , ๐‘“ ๐‘— โŸฉ| 2 . The numbers โŸจ๐‘‡๐‘’ ๐‘˜ , ๐‘“ ๐‘— โŸฉ are the entries of the matrix of ๐‘‡ with respect to the orthonormal bases ๐‘’ 1 , โ€ฆ , ๐‘’ ๐‘› and ๐‘“ 1 , โ€ฆ , ๐‘“ ๐‘š . These numbers depend on the bases, but tr (๐‘‡ โˆ— ๐‘‡) does not depend on a choice of bases. Thus this exercise shows that the sum of the squares of the absolute values of the matrix entries does not depend on which orthonormal bases are used. 12 Suppose ๐‘‰ and ๐‘Š are finite-dimensional inner product spaces. (a) Prove that โŸจ๐‘† , ๐‘‡โŸฉ = tr (๐‘‡ โˆ— ๐‘†) defines an inner product on โ„’ (๐‘‰ , ๐‘Š) . (b) Suppose ๐‘’ 1 , โ€ฆ , ๐‘’ ๐‘› is an orthonormal basis of ๐‘‰ and ๐‘“ 1 , โ€ฆ , ๐‘“ ๐‘š is an or- thonormal basis of ๐‘Š . Show that the inner product on โ„’ (๐‘‰ , ๐‘Š) from (a) is the same as the standard inner product on ๐… ๐‘š๐‘› , where we identify each element of โ„’ (๐‘‰ , ๐‘Š) with its matrix (with respect to the bases just mentioned) and then with an element of ๐… ๐‘š๐‘› . Caution: The norm of a linear map ๐‘‡ โˆˆ โ„’ (๐‘‰ , ๐‘Š) as defined by 7.86 is not the same as the norm that comes from the inner product in ( a ) above. Unless explicitly stated otherwise, always assume that โ€–๐‘‡โ€– refers to the norm as defined by 7.86. The norm that comes from the inner product in ( a ) is called the Frobenius norm or the Hilbertโ€“Schmidt norm . 13 Find ๐‘† , ๐‘‡ โˆˆ โ„’ ( ๐’ซ (๐…)) such that ๐‘†๐‘‡ โˆ’ ๐‘‡๐‘† = ๐ผ . Hint: Make an appropriate modification of the operators in Example 3.9. This exercise shows that additional hypotheses are needed on ๐‘† and ๐‘‡ to extend 8.57 to the setting of infinite-dimensional vector spaces. Linear Algebra Done Right , fourth edition, by Sheldon Axler