Chapter 5 Eigenvalues and Eigenvectors Linear maps from one vector space to another vector space were the objects of study in Chapter 3. Now we begin our investigation of operators, which are linear maps from a vector space to itself. Their study constitutes the most important part of linear algebra. To learn about an operator, we might try restricting it to a smaller subspace. Asking for that restriction to be an operator will lead us to the notion of invariant subspaces. Each one-dimensional invariant subspace arises from a vector that the operator maps into a scalar multiple of the vector. This path will lead us to eigenvectors and eigenvalues. We will then prove one of the most important results in linear algebra: every operator on a finite-dimensional nonzero complex vector space has an eigenvalue. This result will allow us to show that for each operator on a finite-dimensional complex vector space, there is a basis of the vector space with respect to which the matrix of the operator has at least almost half its entries equal to 0 . standing assumptions for this chapter โ€ข ๐… denotes ๐‘ or ๐‚ . โ€ข ๐‘‰ denotes a vector space over ๐… . Hans - Peter Postel CC BY Statue of Leonardo of Pisa ( 1170โ€“1250, approximate dates ) , also known as Fibonacci. Exercise 21 in Section 5D shows how linear algebra can be used to find the explicit formula for the Fibonacci sequence shown on the front cover. Section 5A Invariant Subspaces 133 5A Invariant Subspaces Eigenvalues 5.1 definition: operator A linear map from a vector space to itself is called an operator . Recall that we defined the notation โ„’ (๐‘‰) to mean โ„’ (๐‘‰ , ๐‘‰) . Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . If ๐‘š โ‰ฅ 2 and ๐‘‰ = ๐‘‰ 1 โŠ• โ‹ฏ โŠ• ๐‘‰ ๐‘š , where each ๐‘‰ ๐‘˜ is a nonzero subspace of ๐‘‰ , then to understand the behavior of ๐‘‡ we only need to understand the behavior of each ๐‘‡| ๐‘‰ ๐‘˜ ; here ๐‘‡| ๐‘‰ ๐‘˜ denotes the restriction of ๐‘‡ to the smaller domain ๐‘‰ ๐‘˜ . Dealing with ๐‘‡| ๐‘‰ ๐‘˜ should be easier than dealing with ๐‘‡ because ๐‘‰ ๐‘˜ is a smaller vector space than ๐‘‰ . However, if we intend to apply tools useful in the study of operators (such as taking powers), then we have a problem: ๐‘‡| ๐‘‰ ๐‘˜ may not map ๐‘‰ ๐‘˜ into itself; in other words, ๐‘‡| ๐‘‰ ๐‘˜ may not be an operator on ๐‘‰ ๐‘˜ . Thus we are led to consider only decompositions of ๐‘‰ of the form above in which ๐‘‡ maps each ๐‘‰ ๐‘˜ into itself. Hence we now give a name to subspaces of ๐‘‰ that get mapped into themselves by ๐‘‡ . 5.2 definition: invariant subspace Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . A subspace ๐‘ˆ of ๐‘‰ is called invariant under ๐‘‡ if ๐‘‡๐‘ข โˆˆ ๐‘ˆ for every ๐‘ข โˆˆ ๐‘ˆ . Thus ๐‘ˆ is invariant under ๐‘‡ if ๐‘‡| ๐‘ˆ is an operator on ๐‘ˆ . 5.3 example: subspace invariant under differentiation operator Suppose that ๐‘‡ โˆˆ โ„’ ( ๐’ซ (๐‘)) is defined by ๐‘‡๐‘ = ๐‘ โ€ฒ . Then ๐’ซ 4 (๐‘) , which is a subspace of ๐’ซ (๐‘) , is invariant under ๐‘‡ because if ๐‘ โˆˆ ๐’ซ (๐‘) has degree at most 4 , then ๐‘ โ€ฒ also has degree at most 4 . 5.4 example: four invariant subspaces, not necessarily all different If ๐‘‡ โˆˆ โ„’ (๐‘‰) , then the following subspaces of ๐‘‰ are all invariant under ๐‘‡ . {0} The subspace {0} is invariant under ๐‘‡ because if ๐‘ข โˆˆ {0} , then ๐‘ข = 0 and hence ๐‘‡๐‘ข = 0 โˆˆ {0} . ๐‘‰ The subspace ๐‘‰ is invariant under ๐‘‡ because if ๐‘ข โˆˆ ๐‘‰ , then ๐‘‡๐‘ข โˆˆ ๐‘‰ . null ๐‘‡ The subspace null ๐‘‡ is invariant under ๐‘‡ because if ๐‘ข โˆˆ null ๐‘‡ , then ๐‘‡๐‘ข = 0 , and hence ๐‘‡๐‘ข โˆˆ null ๐‘‡ . range ๐‘‡ The subspace range ๐‘‡ is invariant under ๐‘‡ because if ๐‘ข โˆˆ range ๐‘‡ , then ๐‘‡๐‘ข โˆˆ range ๐‘‡ . Must an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) have any invariant subspaces other than {0} and ๐‘‰ ? Later we will see that this question has an affirmative answer if ๐‘‰ is finite-dimensional and dim ๐‘‰ > 1 (for ๐… = ๐‚ ) or dim ๐‘‰ > 2 (for ๐… = ๐‘) ; see 5.19 and Exercise 29 in Section 5B. The previous example noted that null ๐‘‡ and range ๐‘‡ are invariant under ๐‘‡ . However, these subspaces do not necessarily provide easy answers to the question above about the existence of invariant subspaces other than {0} and ๐‘‰ , because null ๐‘‡ may equal {0} and range ๐‘‡ may equal ๐‘‰ (this happens when ๐‘‡ is invertible). We will return later to a deeper study of invariant subspaces. Now we turn to an investigation of the simplest possible nontrivial invariant subspacesโ€”invariant subspaces of dimension one. Take any ๐‘ฃ โˆˆ ๐‘‰ with ๐‘ฃ โ‰  0 and let ๐‘ˆ equal the set of all scalar multiples of ๐‘ฃ : ๐‘ˆ = {๐œ†๐‘ฃ โˆถ ๐œ† โˆˆ ๐…} = span (๐‘ฃ). Then ๐‘ˆ is a one-dimensional subspace of ๐‘‰ (and every one-dimensional subspace of ๐‘‰ is of this form for an appropriate choice of ๐‘ฃ ). If ๐‘ˆ is invariant under an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) , then ๐‘‡๐‘ฃ โˆˆ ๐‘ˆ , and hence there is a scalar ๐œ† โˆˆ ๐… such that ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ. Conversely, if ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ for some ๐œ† โˆˆ ๐… , then span (๐‘ฃ) is a one-dimensional subspace of ๐‘‰ invariant under ๐‘‡ . The equation ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ , which we have just seen is intimately connected with one-dimensional invariant subspaces, is important enough that the scalars ๐œ† and vectors ๐‘ฃ satisfying it are given special names. 5.5 definition: eigenvalue Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . A number ๐œ† โˆˆ ๐… is called an eigenvalue of ๐‘‡ if there exists ๐‘ฃ โˆˆ ๐‘‰ such that ๐‘ฃ โ‰  0 and ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ . The word eigenvalue is half-German, half-English. The German prefix eigen means โ€œownโ€ in the sense of characterizing an intrinsic property. In the definition above, we require that ๐‘ฃ โ‰  0 because every scalar ๐œ† โˆˆ ๐… satisfies ๐‘‡0 = ๐œ†0 . The comments above show that ๐‘‰ has a one-dimensional subspace invariant under ๐‘‡ if and only if ๐‘‡ has an eigenvalue. 5.6 example: eigenvalue Define an operator ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘ฅ , ๐‘ฆ , ๐‘ง) = (7๐‘ฅ + 3๐‘ง , 3๐‘ฅ + 6๐‘ฆ + 9๐‘ง , โˆ’6๐‘ฆ) for (๐‘ฅ , ๐‘ฆ , ๐‘ง) โˆˆ ๐… 3 . Then ๐‘‡(3 , 1 , โˆ’1) = (18 , 6 , โˆ’6) = 6(3 , 1 , โˆ’1) . Thus 6 is an eigenvalue of ๐‘‡ . The equivalences in the next result, along with many deep results in linear algebra, are valid only in the context of finite-dimensional vector spaces. 5.7 equivalent conditions to be an eigenvalue Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐œ† โˆˆ ๐… . Then the following are equivalent. (a) ๐œ† is an eigenvalue of ๐‘‡ . (b) ๐‘‡ โˆ’ ๐œ†๐ผ is not injective. (c) ๐‘‡ โˆ’ ๐œ†๐ผ is not surjective. Reminder: ๐ผ โˆˆ โ„’ (๐‘‰) is the identity operator. Thus ๐ผ๐‘ฃ = ๐‘ฃ for all ๐‘ฃ โˆˆ ๐‘‰ . (d) ๐‘‡ โˆ’ ๐œ†๐ผ is not invertible. Proof Conditions (a) and (b) are equivalent because the equation ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ is equivalent to the equation (๐‘‡ โˆ’ ๐œ†๐ผ)๐‘ฃ = 0 . Conditions (b), (c), and (d) are equivalent by 3.65. 5.8 definition: eigenvector Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… is an eigenvalue of ๐‘‡ . A vector ๐‘ฃ โˆˆ ๐‘‰ is called an eigenvector of ๐‘‡ corresponding to ๐œ† if ๐‘ฃ โ‰  0 and ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ . In other words, a nonzero vector ๐‘ฃ โˆˆ ๐‘‰ is an eigenvector of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) if and only if ๐‘‡๐‘ฃ is a scalar multiple of ๐‘ฃ . Because ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ if and only if (๐‘‡ โˆ’ ๐œ†๐ผ)๐‘ฃ = 0 , a vector ๐‘ฃ โˆˆ ๐‘‰ with ๐‘ฃ โ‰  0 is an eigenvector of ๐‘‡ corresponding to ๐œ† if and only if ๐‘ฃ โˆˆ null (๐‘‡ โˆ’ ๐œ†๐ผ) . 5.9 example: eigenvalues and eigenvectors Suppose ๐‘‡ โˆˆ โ„’ (๐… 2 ) is defined by ๐‘‡(๐‘ค , ๐‘ง) = (โˆ’๐‘ง , ๐‘ค) . (a) First consider the case ๐… = ๐‘ . Then ๐‘‡ is a counterclockwise rotation by 90 โˆ˜ about the origin in ๐‘ 2 . An operator has an eigenvalue if and only if there exists a nonzero vector in its domain that gets sent by the operator to a scalar multiple of itself. A 90 โˆ˜ counterclockwise rotation of a nonzero vector in ๐‘ 2 cannot equal a scalar multiple of itself. Conclusion: if ๐… = ๐‘ , then ๐‘‡ has no eigenvalues (and thus has no eigenvectors). (b) Now consider the case ๐… = ๐‚ . To find eigenvalues of ๐‘‡ , we must find the scalars ๐œ† such that ๐‘‡(๐‘ค , ๐‘ง) = ๐œ†(๐‘ค , ๐‘ง) has some solution other than ๐‘ค = ๐‘ง = 0 . The equation ๐‘‡(๐‘ค , ๐‘ง) = ๐œ†(๐‘ค , ๐‘ง) is equivalent to the simultaneous equations 5.10 โˆ’๐‘ง = ๐œ†๐‘ค , ๐‘ค = ๐œ†๐‘ง. Substituting the value for ๐‘ค given by the second equation into the first equation gives โˆ’๐‘ง = ๐œ† 2 ๐‘ง. 136 Chapter 5 Eigenvalues and Eigenvectors Now ๐‘ง cannot equal 0 [ otherwise 5.10 implies that ๐‘ค = 0 ; we are looking for solutions to 5.10 such that (๐‘ค , ๐‘ง) is not the 0 vector ] , so the equation above leads to the equation โˆ’1 = ๐œ† 2 . The solutions to this equation are ๐œ† = ๐‘– and ๐œ† = โˆ’๐‘– . You can verify that ๐‘– and โˆ’๐‘– are eigenvalues of ๐‘‡ . Indeed, the eigenvectors corresponding to the eigenvalue ๐‘– are the vectors of the form (๐‘ค , โˆ’๐‘ค๐‘–) , with ๐‘ค โˆˆ ๐‚ and ๐‘ค โ‰  0 . Furthermore, the eigenvectors corresponding to the eigenvalue โˆ’๐‘– are the vectors of the form (๐‘ค , ๐‘ค๐‘–) , with ๐‘ค โˆˆ ๐‚ and ๐‘ค โ‰  0 . In the next proof, we again use the equivalence ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ โŸบ (๐‘‡ โˆ’ ๐œ†๐ผ)๐‘ฃ = 0. 5.11 linearly independent eigenvectors Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then every list of eigenvectors of ๐‘‡ corresponding to distinct eigenvalues of ๐‘‡ is linearly independent. Proof Suppose the desired result is false. Then there exists a smallest positive integer ๐‘š such that there exists a linearly dependent list ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š of eigenvectors of ๐‘‡ corresponding to distinct eigenvalues ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š of ๐‘‡ (note that ๐‘š โ‰ฅ 2 because an eigenvector is, by definition, nonzero). Thus there exist ๐‘Ž 1 , โ€ฆ , ๐‘Ž ๐‘š โˆˆ ๐… , none of which are 0 (because of the minimality of ๐‘š ), such that ๐‘Ž 1 ๐‘ฃ 1 + โ‹ฏ + ๐‘Ž ๐‘š ๐‘ฃ ๐‘š = 0. Apply ๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ to both sides of the equation above, getting ๐‘Ž 1 (๐œ† 1 โˆ’ ๐œ† ๐‘š )๐‘ฃ 1 + โ‹ฏ + ๐‘Ž ๐‘šโˆ’1 (๐œ† ๐‘šโˆ’1 โˆ’ ๐œ† ๐‘š )๐‘ฃ ๐‘šโˆ’1 = 0. Because the eigenvalues ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are distinct, none of the coefficients above equal 0 . Thus ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘šโˆ’1 is a linearly dependent list of ๐‘š โˆ’ 1 eigenvectors of ๐‘‡ corresponding to distinct eigenvalues, contradicting the minimality of ๐‘š . This contradiction completes the proof. The result above leads to a short proof of the result below, which puts an upper bound on the number of distinct eigenvalues that an operator can have. 5.12 operator cannot have more eigenvalues than dimension of vector space Suppose ๐‘‰ is finite-dimensional. Then each operator on ๐‘‰ has at most dim ๐‘‰ distinct eigenvalues. Proof Let ๐‘‡ โˆˆ โ„’ (๐‘‰) . Suppose ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are distinct eigenvalues of ๐‘‡ . Let ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š be corresponding eigenvectors. Then 5.11 implies that the list ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š is linearly independent. Thus ๐‘š โ‰ค dim ๐‘‰ (see 2.22), as desired. Section 5A Invariant Subspaces 137 Polynomials Applied to Operators The main reason that a richer theory exists for operators (which map a vector space into itself) than for more general linear maps is that operators can be raised to powers. In this subsection we define that notion and the concept of applying a polynomial to an operator. This concept will be the key tool that we use in the next section when we prove that every operator on a nonzero finite-dimensional complex vector space has an eigenvalue. If ๐‘‡ is an operator, then ๐‘‡๐‘‡ makes sense (see 3.7) and is also an operator on the same vector space as ๐‘‡ . We usually write ๐‘‡ 2 instead of ๐‘‡๐‘‡ . More generally, we have the following definition of ๐‘‡ ๐‘š . 5.13 notation: ๐‘‡ ๐‘š Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘š is a positive integer. โ€ข ๐‘‡ ๐‘š โˆˆ โ„’ (๐‘‰) is defined by ๐‘‡ ๐‘š = ๐‘‡โ‹ฏ๐‘‡ โŸ ๐‘š times . โ€ข ๐‘‡ 0 is defined to be the identity operator ๐ผ on ๐‘‰ . โ€ข If ๐‘‡ is invertible with inverse ๐‘‡ โˆ’1 , then ๐‘‡ โˆ’๐‘š โˆˆ โ„’ (๐‘‰) is defined by ๐‘‡ โˆ’๐‘š = (๐‘‡ โˆ’1 ) ๐‘š . You should verify that if ๐‘‡ is an operator, then ๐‘‡ ๐‘š ๐‘‡ ๐‘› = ๐‘‡ ๐‘š + ๐‘› and (๐‘‡ ๐‘š ) ๐‘› = ๐‘‡ ๐‘š๐‘› , where ๐‘š and ๐‘› are arbitrary integers if ๐‘‡ is invertible and are nonnegative integers if ๐‘‡ is not invertible. Having defined powers of an operator, we can now define what it means to apply a polynomial to an operator. 5.14 notation: ๐‘(๐‘‡) Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ โˆˆ ๐’ซ (๐…) is a polynomial given by ๐‘(๐‘ง) = ๐‘Ž 0 + ๐‘Ž 1 ๐‘ง + ๐‘Ž 2 ๐‘ง 2 + โ‹ฏ + ๐‘Ž ๐‘š ๐‘ง ๐‘š for all ๐‘ง โˆˆ ๐… . Then ๐‘(๐‘‡) is the operator on ๐‘‰ defined by ๐‘(๐‘‡) = ๐‘Ž 0 ๐ผ + ๐‘Ž 1 ๐‘‡ + ๐‘Ž 2 ๐‘‡ 2 + โ‹ฏ + ๐‘Ž ๐‘š ๐‘‡ ๐‘š . This is a new use of the symbol ๐‘ because we are applying ๐‘ to operators, not just elements of ๐… . The idea here is that to evaluate ๐‘(๐‘‡) , we simply replace ๐‘ง with ๐‘‡ in the expression defining ๐‘ . Note that the constant term ๐‘Ž 0 in ๐‘(๐‘ง) becomes the operator ๐‘Ž 0 ๐ผ ( which is a reasonable choice because ๐‘Ž 0 = ๐‘Ž 0 ๐‘ง 0 and thus we should replace ๐‘Ž 0 with ๐‘Ž 0 ๐‘‡ 0 , which equals ๐‘Ž 0 ๐ผ) . 138 Chapter 5 Eigenvalues and Eigenvectors 5.15 example: a polynomial applied to the differentiation operator Suppose ๐ท โˆˆ โ„’ ( ๐’ซ (๐‘)) is the differentiation operator defined by ๐ท๐‘ž = ๐‘ž โ€ฒ and ๐‘ is the polynomial defined by ๐‘(๐‘ฅ) = 7 โˆ’ 3๐‘ฅ + 5๐‘ฅ 2 . Then ๐‘(๐ท) = 7๐ผ โˆ’ 3๐ท + 5๐ท 2 . Thus (๐‘(๐ท))๐‘ž = 7๐‘ž โˆ’ 3๐‘ž โ€ฒ + 5๐‘ž โ€ณ for every ๐‘ž โˆˆ ๐’ซ (๐‘) . If we fix an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) , then the function from ๐’ซ (๐…) to โ„’ (๐‘‰) given by ๐‘ โ†ฆ ๐‘(๐‘‡) is linear, as you should verify. 5.16 definition: product of polynomials If ๐‘ , ๐‘ž โˆˆ ๐’ซ (๐…) , then ๐‘๐‘ž โˆˆ ๐’ซ (๐…) is the polynomial defined by (๐‘๐‘ž)(๐‘ง) = ๐‘(๐‘ง)๐‘ž(๐‘ง) for all ๐‘ง โˆˆ ๐… . The order does not matter in taking products of polynomials of a single operator, as shown by (b) in the next result. 5.17 multiplicative properties Suppose ๐‘ , ๐‘ž โˆˆ ๐’ซ (๐…) and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then (a) (๐‘๐‘ž)(๐‘‡) = ๐‘(๐‘‡)๐‘ž(๐‘‡) ; (b) ๐‘(๐‘‡)๐‘ž(๐‘‡) = ๐‘ž(๐‘‡)๐‘(๐‘‡) . Informal proof: When a product of polynomials is expanded using the dis- tributive property, it does not matter whether the symbol is ๐‘ง or ๐‘‡ . Proof (a) Suppose ๐‘(๐‘ง) = ๐‘š โˆ‘ ๐‘— = 0 ๐‘Ž ๐‘— ๐‘ง ๐‘— and ๐‘ž(๐‘ง) = ๐‘› โˆ‘ ๐‘˜=0 ๐‘ ๐‘˜ ๐‘ง ๐‘˜ for all ๐‘ง โˆˆ ๐… . Then (๐‘๐‘ž)(๐‘ง) = ๐‘š โˆ‘ ๐‘— = 0 ๐‘› โˆ‘ ๐‘˜=0 ๐‘Ž ๐‘— ๐‘ ๐‘˜ ๐‘ง ๐‘— + ๐‘˜ . Thus (๐‘๐‘ž)(๐‘‡) = ๐‘š โˆ‘ ๐‘— = 0 ๐‘› โˆ‘ ๐‘˜=0 ๐‘Ž ๐‘— ๐‘ ๐‘˜ ๐‘‡ ๐‘— + ๐‘˜ = ( ๐‘š โˆ‘ ๐‘— = 0 ๐‘Ž ๐‘— ๐‘‡ ๐‘— )( ๐‘› โˆ‘ ๐‘˜=0 ๐‘ ๐‘˜ ๐‘‡ ๐‘˜ ) = ๐‘(๐‘‡)๐‘ž(๐‘‡). (b) Using (a) twice, we have ๐‘(๐‘‡)๐‘ž(๐‘‡) = (๐‘๐‘ž)(๐‘‡) = (๐‘ž๐‘)(๐‘‡) = ๐‘ž(๐‘‡)๐‘(๐‘‡) . Section 5A Invariant Subspaces 139 We observed earlier that if ๐‘‡ โˆˆ โ„’ (๐‘‰) , then the subspaces null ๐‘‡ and range ๐‘‡ are invariant under ๐‘‡ (see 5.4). Now we show that the null space and the range of every polynomial of ๐‘‡ are also invariant under ๐‘‡ . 5.18 null space and range of ๐‘(๐‘‡) are invariant under ๐‘‡ Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ โˆˆ ๐’ซ (๐…) . Then null ๐‘(๐‘‡) and range ๐‘(๐‘‡) are invariant under ๐‘‡ . Proof Suppose ๐‘ข โˆˆ null ๐‘(๐‘‡) . Then ๐‘(๐‘‡)๐‘ข = 0 . Thus (๐‘(๐‘‡))(๐‘‡๐‘ข) = ๐‘‡(๐‘(๐‘‡)๐‘ข) = ๐‘‡(0) = 0. Hence ๐‘‡๐‘ข โˆˆ null ๐‘(๐‘‡) . Thus null ๐‘(๐‘‡) is invariant under ๐‘‡ , as desired. Suppose ๐‘ข โˆˆ range ๐‘(๐‘‡) . Then there exists ๐‘ฃ โˆˆ ๐‘‰ such that ๐‘ข = ๐‘(๐‘‡)๐‘ฃ . Thus ๐‘‡๐‘ข = ๐‘‡(๐‘(๐‘‡)๐‘ฃ) = ๐‘(๐‘‡)(๐‘‡๐‘ฃ). Hence ๐‘‡๐‘ข โˆˆ range ๐‘(๐‘‡) . Thus range ๐‘(๐‘‡) is invariant under ๐‘‡ , as desired. Exercises 5A 1 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ˆ is a subspace of ๐‘‰ . (a) Prove that if ๐‘ˆ โІ null ๐‘‡ , then ๐‘ˆ is invariant under ๐‘‡ . (b) Prove that if range ๐‘‡ โІ ๐‘ˆ , then ๐‘ˆ is invariant under ๐‘‡ . 2 Suppose that ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘‰ 1 , โ€ฆ , ๐‘‰ ๐‘š are subspaces of ๐‘‰ invariant under ๐‘‡ . Prove that ๐‘‰ 1 + โ‹ฏ + ๐‘‰ ๐‘š is invariant under ๐‘‡ . 3 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that the intersection of every collection of subspaces of ๐‘‰ invariant under ๐‘‡ is invariant under ๐‘‡ . 4 Prove or give a counterexample: If ๐‘‰ is finite-dimensional and ๐‘ˆ is a sub- space of ๐‘‰ that is invariant under every operator on ๐‘‰ , then ๐‘ˆ = {0} or ๐‘ˆ = ๐‘‰ . 5 Suppose ๐‘‡ โˆˆ โ„’ (๐‘ 2 ) is defined by ๐‘‡(๐‘ฅ , ๐‘ฆ) = (โˆ’3๐‘ฆ , ๐‘ฅ) . Find the eigenvalues of ๐‘‡ . 6 Define ๐‘‡ โˆˆ โ„’ (๐… 2 ) by ๐‘‡(๐‘ค , ๐‘ง) = (๐‘ง , ๐‘ค) . Find all eigenvalues and eigenvec- tors of ๐‘‡ . 7 Define ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (2๐‘ง 2 , 0 , 5๐‘ง 3 ) . Find all eigenvalues and eigenvectors of ๐‘‡ . 8 Suppose ๐‘ƒ โˆˆ โ„’ (๐‘‰) is such that ๐‘ƒ 2 = ๐‘ƒ . Prove that if ๐œ† is an eigenvalue of ๐‘ƒ , then ๐œ† = 0 or ๐œ† = 1 . 140 Chapter 5 Eigenvalues and Eigenvectors 9 Define ๐‘‡ โˆถ ๐’ซ (๐‘) โ†’ ๐’ซ (๐‘) by ๐‘‡๐‘ = ๐‘ โ€ฒ . Find all eigenvalues and eigenvectors of ๐‘‡ . 10 Define ๐‘‡ โˆˆ โ„’ ( ๐’ซ 4 (๐‘)) by (๐‘‡๐‘)(๐‘ฅ) = ๐‘ฅ๐‘ โ€ฒ (๐‘ฅ) for all ๐‘ฅ โˆˆ ๐‘ . Find all eigenval- ues and eigenvectors of ๐‘‡ . 11 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐›ผ โˆˆ ๐… . Prove that there ex- ists ๐›ฟ > 0 such that ๐‘‡โˆ’๐œ†๐ผ is invertible for all ๐œ† โˆˆ ๐… such that 0 < |๐›ผ โˆ’ ๐œ†| < ๐›ฟ . 12 Suppose ๐‘‰ = ๐‘ˆ โŠ• ๐‘Š , where ๐‘ˆ and ๐‘Š are nonzero subspaces of ๐‘‰ . Define ๐‘ƒ โˆˆ โ„’ (๐‘‰) by ๐‘ƒ(๐‘ข + ๐‘ค) = ๐‘ข for each ๐‘ข โˆˆ ๐‘ˆ and each ๐‘ค โˆˆ ๐‘Š . Find all eigenvalues and eigenvectors of ๐‘ƒ . 13 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Suppose ๐‘† โˆˆ โ„’ (๐‘‰) is invertible. (a) Prove that ๐‘‡ and ๐‘† โˆ’1 ๐‘‡๐‘† have the same eigenvalues. (b) What is the relationship between the eigenvectors of ๐‘‡ and the eigenvec- tors of ๐‘† โˆ’1 ๐‘‡๐‘† ? 14 Give an example of an operator on ๐‘ 4 that has no (real) eigenvalues. 15 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐œ† โˆˆ ๐… . Show that ๐œ† is an eigenvalue of ๐‘‡ if and only if ๐œ† is an eigenvalue of the dual operator ๐‘‡ โ€ฒ โˆˆ โ„’ (๐‘‰ โ€ฒ ) . 16 Suppose ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that if ๐œ† is an eigenvalue of ๐‘‡ , then |๐œ†| โ‰ค ๐‘› max {โˆฃ โ„ณ (๐‘‡) ๐‘— , ๐‘˜ โˆฃ โˆถ 1 โ‰ค ๐‘— , ๐‘˜ โ‰ค ๐‘›} , where โ„ณ (๐‘‡) ๐‘— , ๐‘˜ denotes the entry in row ๐‘— , column ๐‘˜ of the matrix of ๐‘‡ with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› . See Exercise 19 in Section 6A for a different bound on |๐œ†| . 17 Suppose ๐… = ๐‘ , ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐œ† โˆˆ ๐‘ . Prove that ๐œ† is an eigenvalue of ๐‘‡ if and only if ๐œ† is an eigenvalue of the complexification ๐‘‡ ๐‚ . See Exercise 33 in Section 3B for the definition of ๐‘‡ ๐‚ . 18 Suppose ๐… = ๐‘ , ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐œ† โˆˆ ๐‚ . Prove that ๐œ† is an eigenvalue of the complexification ๐‘‡ ๐‚ if and only if ๐œ† is an eigenvalue of ๐‘‡ ๐‚ . 19 Show that the forward shift operator ๐‘‡ โˆˆ โ„’ (๐… โˆž ) defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , โ€ฆ ) = (0 , ๐‘ง 1 , ๐‘ง 2 , โ€ฆ ) has no eigenvalues. 20 Define the backward shift operator ๐‘† โˆˆ โ„’ (๐… โˆž ) by ๐‘†(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , โ€ฆ ) = (๐‘ง 2 , ๐‘ง 3 , โ€ฆ ). (a) Show that every element of ๐… is an eigenvalue of ๐‘† . (b) Find all eigenvectors of ๐‘† . Section 5A Invariant Subspaces โ‰ˆ 100 โˆš 2 21 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible. (a) Suppose ๐œ† โˆˆ ๐… with ๐œ† โ‰  0 . Prove that ๐œ† is an eigenvalue of ๐‘‡ if and only if 1๐œ† is an eigenvalue of ๐‘‡ โˆ’1 . (b) Prove that ๐‘‡ and ๐‘‡ โˆ’1 have the same eigenvectors. 22 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and there exist nonzero vectors ๐‘ข and ๐‘ค in ๐‘‰ such that ๐‘‡๐‘ข = 3๐‘ค and ๐‘‡๐‘ค = 3๐‘ข. Prove that 3 or โˆ’3 is an eigenvalue of ๐‘‡ . 23 Suppose ๐‘‰ is finite-dimensional and ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘†๐‘‡ and ๐‘‡๐‘† have the same eigenvalues. 24 Suppose ๐ด is an ๐‘› -by- ๐‘› matrix with entries in ๐… . Define ๐‘‡ โˆˆ โ„’ (๐… ๐‘› ) by ๐‘‡๐‘ฅ = ๐ด๐‘ฅ , where elements of ๐… ๐‘› are thought of as ๐‘› -by- 1 column vectors. (a) Suppose the sum of the entries in each row of ๐ด equals 1 . Prove that 1 is an eigenvalue of ๐‘‡ . (b) Suppose the sum of the entries in each column of ๐ด equals 1 . Prove that 1 is an eigenvalue of ๐‘‡ . 25 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ข , ๐‘ค are eigenvectors of ๐‘‡ such that ๐‘ข + ๐‘ค is also an eigenvector of ๐‘‡ . Prove that ๐‘ข and ๐‘ค are eigenvectors of ๐‘‡ corresponding to the same eigenvalue. 26 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that every nonzero vector in ๐‘‰ is an eigenvector of ๐‘‡ . Prove that ๐‘‡ is a scalar multiple of the identity operator. 27 Suppose that ๐‘‰ is finite-dimensional and ๐‘˜ โˆˆ {1 , โ€ฆ , dim ๐‘‰ โˆ’ 1} . Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that every subspace of ๐‘‰ of dimension ๐‘˜ is invariant under ๐‘‡ . Prove that ๐‘‡ is a scalar multiple of the identity operator. 28 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ has at most 1 + dim range ๐‘‡ distinct eigenvalues. 29 Suppose ๐‘‡ โˆˆ โ„’ (๐‘ 3 ) and โˆ’4 , 5 , and โˆš 7 are eigenvalues of ๐‘‡ . Prove that there exists ๐‘ฅ โˆˆ ๐‘ 3 such that ๐‘‡๐‘ฅ โˆ’ 9๐‘ฅ = (โˆ’4 , 5 , โˆš 7) . 30 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and (๐‘‡ โˆ’ 2๐ผ)(๐‘‡ โˆ’ 3๐ผ)(๐‘‡ โˆ’ 4๐ผ) = 0 . Suppose ๐œ† is an eigenvalue of ๐‘‡ . Prove that ๐œ† = 2 or ๐œ† = 3 or ๐œ† = 4 . 31 Give an example of ๐‘‡ โˆˆ โ„’ (๐‘ 2 ) such that ๐‘‡ 4 = โˆ’๐ผ . 32 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) has no eigenvalues and ๐‘‡ 4 = ๐ผ . Prove that ๐‘‡ 2 = โˆ’๐ผ . 33 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘š is a positive integer. (a) Prove that ๐‘‡ is injective if and only if ๐‘‡ ๐‘š is injective. (b) Prove that ๐‘‡ is surjective if and only if ๐‘‡ ๐‘š is surjective. 142 Chapter 5 Eigenvalues and Eigenvectors 34 Suppose ๐‘‰ is finite-dimensional and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š โˆˆ ๐‘‰ . Prove that the list ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š is linearly independent if and only if there exists ๐‘‡ โˆˆ โ„’ (๐‘‰) such that ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘š are eigenvectors of ๐‘‡ corresponding to distinct eigenvalues. 35 Suppose that ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› is a list of distinct real numbers. Prove that the list ๐‘’ ๐œ† 1 ๐‘ฅ , โ€ฆ , ๐‘’ ๐œ† ๐‘› ๐‘ฅ is linearly independent in the vector space of real-valued functions on ๐‘ . Hint: Let ๐‘‰ = span (๐‘’ ๐œ† 1 ๐‘ฅ , โ€ฆ , ๐‘’ ๐œ† ๐‘› ๐‘ฅ ) , and define an operator ๐ท โˆˆ โ„’ (๐‘‰) by ๐ท ๐‘“ = ๐‘“ โ€ฒ . Find eigenvalues and eigenvectors of ๐ท . 36 Suppose that ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› is a list of distinct positive numbers. Prove that the list cos (๐œ† 1 ๐‘ฅ) , โ€ฆ , cos (๐œ† ๐‘› ๐‘ฅ) is linearly independent in the vector space of real-valued functions on ๐‘ . 37 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Define ๐’œ โˆˆ โ„’ ( โ„’ (๐‘‰)) by ๐’œ (๐‘†) = ๐‘‡๐‘† for each ๐‘† โˆˆ โ„’ (๐‘‰) . Prove that the set of eigenvalues of ๐‘‡ equals the set of eigenvalues of ๐’œ . 38 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ˆ is a subspace of ๐‘‰ invariant under ๐‘‡ . The quotient operator ๐‘‡/๐‘ˆ โˆˆ โ„’ (๐‘‰/๐‘ˆ) is defined by (๐‘‡/๐‘ˆ)(๐‘ฃ + ๐‘ˆ) = ๐‘‡๐‘ฃ + ๐‘ˆ for each ๐‘ฃ โˆˆ ๐‘‰ . (a) Show that the definition of ๐‘‡/๐‘ˆ makes sense (which requires using the condition that ๐‘ˆ is invariant under ๐‘‡ ) and show that ๐‘‡/๐‘ˆ is an operator on ๐‘‰/๐‘ˆ . (b) Show that each eigenvalue of ๐‘‡/๐‘ˆ is an eigenvalue of ๐‘‡ . 39 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ has an eigen- value if and only if there exists a subspace of ๐‘‰ of dimension dim ๐‘‰ โˆ’ 1 that is invariant under ๐‘‡ . 40 Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘† is invertible. Suppose ๐‘ โˆˆ ๐’ซ (๐…) is a polynomial. Prove that ๐‘(๐‘†๐‘‡๐‘† โˆ’1 ) = ๐‘†๐‘(๐‘‡)๐‘† โˆ’1 . 41 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ˆ is a subspace of ๐‘‰ invariant under ๐‘‡ . Prove that ๐‘ˆ is invariant under ๐‘(๐‘‡) for every polynomial ๐‘ โˆˆ ๐’ซ (๐…) . 42 Define ๐‘‡ โˆˆ โ„’ (๐… ๐‘› ) by ๐‘‡(๐‘ฅ 1 , ๐‘ฅ 2 , ๐‘ฅ 3 , โ€ฆ , ๐‘ฅ ๐‘› ) = (๐‘ฅ 1 , 2๐‘ฅ 2 , 3๐‘ฅ 3 , โ€ฆ , ๐‘›๐‘ฅ ๐‘› ) . (a) Find all eigenvalues and eigenvectors of ๐‘‡ . (b) Find all subspaces of ๐… ๐‘› that are invariant under ๐‘‡ . 43 Suppose that ๐‘‰ is finite-dimensional, dim ๐‘‰ > 1 , and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that {๐‘(๐‘‡) โˆถ ๐‘ โˆˆ ๐’ซ (๐…)} โ‰  โ„’ (๐‘‰) . Section 5B The Minimal Polynomial 143 5B The Minimal Polynomial Existence of Eigenvalues on Complex Vector Spaces Now we come to one of the central results about operators on finite-dimensional complex vector spaces. 5.19 existence of eigenvalues Every operator on a finite-dimensional nonzero complex vector space has an eigenvalue. Proof Suppose ๐‘‰ is a finite-dimensional complex vector space of dimension ๐‘› > 0 and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Choose ๐‘ฃ โˆˆ ๐‘‰ with ๐‘ฃ โ‰  0 . Then ๐‘ฃ , ๐‘‡๐‘ฃ , ๐‘‡ 2 ๐‘ฃ , โ€ฆ , ๐‘‡ ๐‘› ๐‘ฃ is not linearly independent, because ๐‘‰ has dimension ๐‘› and this list has length ๐‘› + 1 . Hence some linear combination (with not all the coefficients equal to 0 ) of the vectors above equals 0 . Thus there exists a nonconstant polynomial ๐‘ of smallest degree such that ๐‘(๐‘‡)๐‘ฃ = 0. By the first version of the fundamental theorem of algebra (see 4.12), there exists ๐œ† โˆˆ ๐‚ such that ๐‘(๐œ†) = 0 . Hence there exists a polynomial ๐‘ž โˆˆ ๐’ซ (๐‚) such that ๐‘(๐‘ง) = (๐‘ง โˆ’ ๐œ†)๐‘ž(๐‘ง) for every ๐‘ง โˆˆ ๐‚ (see 4.6). This implies (using 5.17) that 0 = ๐‘(๐‘‡)๐‘ฃ = (๐‘‡ โˆ’ ๐œ†๐ผ)(๐‘ž(๐‘‡)๐‘ฃ). Because ๐‘ž has smaller degree than ๐‘ , we know that ๐‘ž(๐‘‡)๐‘ฃ โ‰  0 . Thus the equation above implies that ๐œ† is an eigenvalue of ๐‘‡ with eigenvector ๐‘ž(๐‘‡)๐‘ฃ . The proof above makes crucial use of the fundamental theorem of algebra. The comment following Exercise 16 helps explain why the fundamental theorem of algebra is so tightly connected to the result above. The hypothesis in the result above that ๐… = ๐‚ cannot be replaced with the hypothesis that ๐… = ๐‘ , as shown by Example 5.9. The next example shows that the finite-dimensional hypothesis in the result above also cannot be deleted. 5.20 example: an operator on a complex vector space with no eigenvalues Define ๐‘‡ โˆˆ โ„’ ( ๐’ซ (๐‚)) by (๐‘‡๐‘)(๐‘ง) = ๐‘ง๐‘(๐‘ง) . If ๐‘ โˆˆ ๐’ซ (๐‚) is a nonzero polynomial, then the degree of ๐‘‡๐‘ is one more than the degree of ๐‘ , and thus ๐‘‡๐‘ cannot equal a scalar multiple of ๐‘ . Hence ๐‘‡ has no eigenvalues. Because ๐’ซ (๐‚) is infinite-dimensional, this example does not contradict the result above. 144 Chapter 5 Eigenvalues and Eigenvectors Eigenvalues and the Minimal Polynomial In this subsection we introduce an important polynomial associated with each operator. We begin with the following definition. 5.21 definition: monic polynomial A monic polynomial is a polynomial whose highest-degree coefficient equals 1 . For example, the polynomial 2 + 9๐‘ง 2 + ๐‘ง 7 is a monic polynomial of degree 7 . 5.22 existence, uniqueness, and degree of minimal polynomial Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then there is a unique monic polynomial ๐‘ โˆˆ ๐’ซ (๐…) of smallest degree such that ๐‘(๐‘‡) = 0 . Furthermore, deg ๐‘ โ‰ค dim ๐‘‰ . Proof If dim ๐‘‰ = 0 , then ๐ผ is the zero operator on ๐‘‰ and thus we take ๐‘ to be the constant polynomial 1 . Now use induction on dim ๐‘‰ . Thus assume that dim ๐‘‰ > 0 and that the desired result is true for all operators on all vector spaces of smaller dimension. Let ๐‘ข โˆˆ ๐‘‰ be such that ๐‘ข โ‰  0 . The list ๐‘ข , ๐‘‡๐‘ข , โ€ฆ , ๐‘‡ dim ๐‘‰ ๐‘ข has length 1 + dim ๐‘‰ and thus is linearly dependent. By the linear dependence lemma (2.19), there is a smallest positive integer ๐‘š โ‰ค dim ๐‘‰ such that ๐‘‡ ๐‘š ๐‘ข is a linear combination of ๐‘ข , ๐‘‡๐‘ข , โ€ฆ , ๐‘‡ ๐‘šโˆ’1 ๐‘ข . Thus there exist scalars ๐‘ 0 , ๐‘ 1 , ๐‘ 2 , โ€ฆ , ๐‘ ๐‘šโˆ’1 โˆˆ ๐… such that 5.23 ๐‘ 0 ๐‘ข + ๐‘ 1 ๐‘‡๐‘ข + โ‹ฏ + ๐‘ ๐‘šโˆ’1 ๐‘‡ ๐‘šโˆ’1 ๐‘ข + ๐‘‡ ๐‘š ๐‘ข = 0. Define a monic polynomial ๐‘ž โˆˆ ๐’ซ ๐‘š (๐…) by ๐‘ž(๐‘ง) = ๐‘ 0 + ๐‘ 1 ๐‘ง + โ‹ฏ + ๐‘ ๐‘šโˆ’1 ๐‘ง ๐‘šโˆ’1 + ๐‘ง ๐‘š . Then 5.23 implies that ๐‘ž(๐‘‡)๐‘ข = 0 . If ๐‘˜ is a nonnegative integer, then ๐‘ž(๐‘‡)(๐‘‡ ๐‘˜ ๐‘ข) = ๐‘‡ ๐‘˜ (๐‘ž(๐‘‡)๐‘ข) = ๐‘‡ ๐‘˜ (0) = 0. The linear dependence lemma (2.19) shows that ๐‘ข , ๐‘‡๐‘ข , โ€ฆ , ๐‘‡ ๐‘šโˆ’1 ๐‘ข is linearly inde- pendent. Thus the equation above implies that dim null ๐‘ž(๐‘‡) โ‰ฅ ๐‘š . Hence dim range ๐‘ž(๐‘‡) = dim ๐‘‰ โˆ’ dim null ๐‘ž(๐‘‡) โ‰ค dim ๐‘‰ โˆ’ ๐‘š. Because range ๐‘ž(๐‘‡) is invariant under ๐‘‡ (by 5.18), we can apply our induction hypothesis to the operator ๐‘‡| range ๐‘ž(๐‘‡) on the vector space range ๐‘ž(๐‘‡) . Thus there is a monic polynomial ๐‘  โˆˆ ๐’ซ (๐…) with deg ๐‘  โ‰ค dim ๐‘‰ โˆ’ ๐‘š and ๐‘ (๐‘‡| range ๐‘ž(๐‘‡) ) = 0. Hence for all ๐‘ฃ โˆˆ ๐‘‰ we have ((๐‘ ๐‘ž)(๐‘‡))(๐‘ฃ) = ๐‘ (๐‘‡)(๐‘ž(๐‘‡)๐‘ฃ) = 0 because ๐‘ž(๐‘‡)๐‘ฃ โˆˆ range ๐‘ž(๐‘‡) and ๐‘ (๐‘‡)| range ๐‘ž(๐‘‡) = ๐‘ (๐‘‡| range ๐‘ž(๐‘‡) ) = 0 . Thus ๐‘ ๐‘ž is a monic polynomial such that deg ๐‘ ๐‘ž โ‰ค dim ๐‘‰ and (๐‘ ๐‘ž)(๐‘‡) = 0 . Section 5B The Minimal Polynomial 145 The paragraph above shows that there is a monic polynomial of degree at most dim ๐‘‰ that when applied to ๐‘‡ gives the 0 operator. Thus there is a monic polynomial of smallest degree with this property, completing the existence part of this result. Let ๐‘ โˆˆ ๐’ซ (๐…) be a monic polynomial of smallest degree such that ๐‘(๐‘‡) = 0 . To prove the uniqueness part of the result, suppose ๐‘Ÿ โˆˆ ๐’ซ (๐…) is a monic poly-nomial of the same degree as ๐‘ and ๐‘Ÿ(๐‘‡) = 0 . Then (๐‘ โˆ’ ๐‘Ÿ)(๐‘‡) = 0 and also deg (๐‘ โˆ’ ๐‘Ÿ) < deg ๐‘ . If ๐‘ โˆ’ ๐‘Ÿ were not equal to 0 , then we could divide ๐‘ โˆ’ ๐‘Ÿ by the coefficient of the highest-order term in ๐‘ โˆ’ ๐‘Ÿ to get a monic polynomial (of smaller degree than ๐‘ ) that when applied to ๐‘‡ gives the 0 operator. Thus ๐‘โˆ’๐‘Ÿ = 0 , as desired. The previous result justifies the following definition. 5.24 definition: minimal polynomial Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then the minimal polynomial of ๐‘‡ is the unique monic polynomial ๐‘ โˆˆ ๐’ซ (๐…) of smallest degree such that ๐‘(๐‘‡) = 0 . To compute the minimal polynomial of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) , we need to find the smallest positive integer ๐‘š such that the equation ๐‘ 0 ๐ผ + ๐‘ 1 ๐‘‡ + โ‹ฏ + ๐‘ ๐‘šโˆ’1 ๐‘‡ ๐‘šโˆ’1 = โˆ’๐‘‡ ๐‘š has a solution ๐‘ 0 , ๐‘ 1 , โ€ฆ , ๐‘ ๐‘šโˆ’1 โˆˆ ๐… . If we pick a basis of ๐‘‰ and replace ๐‘‡ in the equation above with the matrix of ๐‘‡ , then the equation above can be thought of as a system of ( dim ๐‘‰) 2 linear equations in the ๐‘š unknowns ๐‘ 0 , ๐‘ 1 , โ€ฆ , ๐‘ ๐‘šโˆ’1 โˆˆ ๐… . Gaussian elimination or another fast method of solving systems of linear equations can tell us whether a solution exists, testing successive values ๐‘š = 1 , 2 , โ€ฆ until a solution exists. By 5.22, a solution exists for some smallest positive integer ๐‘š โ‰ค dim ๐‘‰ . The minimal polynomial of ๐‘‡ is then ๐‘ 0 + ๐‘ 1 ๐‘ง + โ‹ฏ + ๐‘ ๐‘šโˆ’1 ๐‘ง ๐‘šโˆ’1 + ๐‘ง ๐‘š . Even faster (usually), pick ๐‘ฃ โˆˆ ๐‘‰ and consider the equation 5.25 ๐‘ 0 ๐‘ฃ + ๐‘ 1 ๐‘‡๐‘ฃ + โ‹ฏ + ๐‘ dim ๐‘‰โˆ’1 ๐‘‡ dim ๐‘‰โˆ’1 ๐‘ฃ = โˆ’๐‘‡ dim ๐‘‰ ๐‘ฃ. Use a basis of ๐‘‰ to convert the equation above to a system of dim ๐‘‰ linear equations in dim ๐‘‰ unknowns ๐‘ 0 , ๐‘ 1 , โ€ฆ , ๐‘ dim ๐‘‰โˆ’1 . If this system of equations has a unique solution ๐‘ 0 , ๐‘ 1 , โ€ฆ , ๐‘ dim ๐‘‰โˆ’1 (as happens most of the time), then the scalars ๐‘ 0 , ๐‘ 1 , โ€ฆ , ๐‘ dim ๐‘‰โˆ’1 , 1 are the coefficients of the minimal polynomial of ๐‘‡ (because 5.22 states that the degree of the minimal polynomial is at most dim ๐‘‰ ). These estimates are based on testing millions of random matrices. Consider operators on ๐‘ 4 (thought of as 4 -by- 4 matrices with respect to the standard basis), and take ๐‘ฃ = (1 , 0 , 0 , 0) in the paragraph above. The faster method described above works on over 99.8% of the 4 -by- 4 matrices with integer entries in the interval [โˆ’10 , 10] and on over 99.999% of the 4 -by- 4 matrices with integer entries in [โˆ’100 , 100] . 146 Chapter 5 Eigenvalues and Eigenvectors The next example illustrates the faster procedure discussed above. 5.26 example: minimal polynomial of an operator on ๐… 5 Suppose ๐‘‡ โˆˆ โ„’ (๐… 5 ) and โ„ณ (๐‘‡) = โŽ› โŽœ โŽœ โŽœ โŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽ 0 0 0 0 โˆ’3 1 0 0 0 6 0 1 0 0 0 0 0 1 0 0 0 0 0 1 0 โŽž โŽŸ โŽŸ โŽŸ โŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽ  with respect to the standard basis ๐‘’ 1 , ๐‘’ 2 , ๐‘’ 3 , ๐‘’ 4 , ๐‘’ 5 . Taking ๐‘ฃ = ๐‘’ 1 for 5.25, we have ๐‘‡๐‘’ 1 = ๐‘’ 2 , ๐‘‡ 4 ๐‘’ 1 = ๐‘‡(๐‘‡ 3 ๐‘’ 1 ) = ๐‘‡๐‘’ 4 = ๐‘’ 5 , ๐‘‡ 2 ๐‘’ 1 = ๐‘‡(๐‘‡๐‘’ 1 ) = ๐‘‡๐‘’ 2 = ๐‘’ 3 , ๐‘‡ 5 ๐‘’ 1 = ๐‘‡(๐‘‡ 4 ๐‘’ 1 ) = ๐‘‡๐‘’ 5 = โˆ’3๐‘’ 1 + 6๐‘’ 2 . ๐‘‡ 3 ๐‘’ 1 = ๐‘‡(๐‘‡ 2 ๐‘’ 1 ) = ๐‘‡๐‘’ 3 = ๐‘’ 4 , Thus 3๐‘’ 1 โˆ’ 6๐‘‡๐‘’ 1 = โˆ’๐‘‡ 5 ๐‘’ 1 . The list ๐‘’ 1 , ๐‘‡๐‘’ 1 , ๐‘‡ 2 ๐‘’ 1 , ๐‘‡ 3 ๐‘’ 1 , ๐‘‡ 4 ๐‘’ 1 , which equals the list ๐‘’ 1 , ๐‘’ 2 , ๐‘’ 3 , ๐‘’ 4 , ๐‘’ 5 , is linearly independent, so no other linear combination of this list equals โˆ’๐‘‡ 5 ๐‘’ 1 . Hence the minimal polynomial of ๐‘‡ is 3 โˆ’ 6๐‘ง + ๐‘ง 5 . Recall that by definition, eigenvalues of operators on ๐‘‰ and zeros of polynomials in ๐’ซ (๐…) must be elements of ๐… . In particular, if ๐… = ๐‘ , then eigenvalues and zeros must be real numbers. 5.27 eigenvalues are the zeros of the minimal polynomial Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . (a) The zeros of the minimal polynomial of ๐‘‡ are the eigenvalues of ๐‘‡ . (b) If ๐‘‰ is a complex vector space, then the minimal polynomial of ๐‘‡ has the form (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) , where ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š is a list of all eigenvalues of ๐‘‡ , possibly with repetitions. Proof Let ๐‘ be the minimal polynomial of ๐‘‡ . (a) First suppose ๐œ† โˆˆ ๐… is a zero of ๐‘ . Then ๐‘ can be written in the form ๐‘(๐‘ง) = (๐‘ง โˆ’ ๐œ†)๐‘ž(๐‘ง) , where ๐‘ž is a monic polynomial with coefficients in ๐… (see 4.6). Because ๐‘(๐‘‡) = 0 , we have 0 = (๐‘‡ โˆ’ ๐œ†๐ผ)(๐‘ž(๐‘‡)๐‘ฃ) for all ๐‘ฃ โˆˆ ๐‘‰ . Because deg ๐‘ž = ( deg ๐‘) โˆ’ 1 and ๐‘ is the minimal polynomial of ๐‘‡ , there exists at least one vector ๐‘ฃ โˆˆ ๐‘‰ such that ๐‘ž(๐‘‡)๐‘ฃ โ‰  0 . The equation above thus implies that ๐œ† is an eigenvalue of ๐‘‡ , as desired. Section 5B The Minimal Polynomial 147 To prove that every eigenvalue of ๐‘‡ is a zero of ๐‘ , now suppose ๐œ† โˆˆ ๐… is an eigenvalue of ๐‘‡ . Thus there exists ๐‘ฃ โˆˆ ๐‘‰ with ๐‘ฃ โ‰  0 such that ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ . Repeated applications of ๐‘‡ to both sides of this equation show that ๐‘‡ ๐‘˜ ๐‘ฃ = ๐œ† ๐‘˜ ๐‘ฃ for every nonnegative integer ๐‘˜ . Thus ๐‘(๐‘‡)๐‘ฃ = ๐‘(๐œ†)๐‘ฃ. Because ๐‘ is the minimal polynomial of ๐‘‡ , we have ๐‘(๐‘‡)๐‘ฃ = 0 . Hence the equation above implies that ๐‘(๐œ†) = 0 . Thus ๐œ† is a zero of ๐‘ , as desired. (b) To get the desired result, use (a) and the second version of the fundamental theorem of algebra (see 4.13). A nonzero polynomial has at most as many distinct zeros as its degree (see 4.8). Thus (a) of the previous result, along with the result that the minimal polynomial of an operator on ๐‘‰ has degree at most dim ๐‘‰ , gives an alternative proof of 5.12, which states that an operator on ๐‘‰ has at most dim ๐‘‰ distinct eigenvalues. Every monic polynomial is the minimal polynomial of some operator, as shown by Exercise 16, which generalizes Example 5.26. Thus 5.27(a) shows that finding exact expressions for the eigenvalues of an operator is equivalent to the problem of finding exact expressions for the zeros of a polynomial (and thus is not possible for some operators). 5.28 example: An operator whose eigenvalues cannot be found exactly Let ๐‘‡ โˆˆ โ„’ (๐‚ 5 ) be the operator defined by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 , ๐‘ง 5 ) = (โˆ’3๐‘ง 5 , ๐‘ง 1 + 6๐‘ง 5 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 ). The matrix of ๐‘‡ with respect to the standard basis of ๐‚ 5 is the 5 -by- 5 matrix in Example 5.26. As we showed in that example, the minimal polynomial of ๐‘‡ is the polynomial 3 โˆ’ 6๐‘ง + ๐‘ง 5 . No zero of the polynomial above can be expressed using rational numbers, roots of rational numbers, and the usual rules of arithmetic (a proof of this would take us considerably beyond linear algebra). Because the zeros of the polynomial above are the eigenvalues of ๐‘‡ [ by 5.27(a) ] , we cannot find an exact expression for any eigenvalue of ๐‘‡ in any familiar form. Numeric techniques, which we will not discuss here, show that the zeros of the polynomial above, and thus the eigenvalues of ๐‘‡ , are approximately the following five complex numbers: โˆ’1.67 , 0.51 , 1.40 , โˆ’0.12 + 1.59๐‘– , โˆ’0.12 โˆ’ 1.59๐‘–. Note that the two nonreal zeros of this polynomial are complex conjugates of each other, as we expect for a polynomial with real coefficients (see 4.14). 148 Chapter 5 Eigenvalues and Eigenvectors The next result completely characterizes the polynomials that when applied to an operator give the 0 operator. 5.29 ๐‘ž(๐‘‡) = 0 โŸบ ๐‘ž is a polynomial multiple of the minimal polynomial Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ž โˆˆ ๐’ซ (๐…) . Then ๐‘ž(๐‘‡) = 0 if and only if ๐‘ž is a polynomial multiple of the minimal polynomial of ๐‘‡ . Proof Let ๐‘ denote the minimal polynomial of ๐‘‡ . First suppose ๐‘ž(๐‘‡) = 0 . By the division algorithm for polynomials (4.9), there exist polynomials ๐‘  , ๐‘Ÿ โˆˆ ๐’ซ (๐…) such that 5.30 ๐‘ž = ๐‘๐‘  + ๐‘Ÿ and deg ๐‘Ÿ < deg ๐‘ . We have 0 = ๐‘ž(๐‘‡) = ๐‘(๐‘‡)๐‘ (๐‘‡) + ๐‘Ÿ(๐‘‡) = ๐‘Ÿ(๐‘‡). The equation above implies that ๐‘Ÿ = 0 (otherwise, dividing ๐‘Ÿ by its highest-degree coefficient would produce a monic polynomial that when applied to ๐‘‡ gives 0 ; this polynomial would have a smaller degree than the minimal polynomial, which would be a contradiction). Thus 5.30 becomes the equation ๐‘ž = ๐‘๐‘  . Hence ๐‘ž is a polynomial multiple of ๐‘ , as desired. To prove the other direction, now suppose ๐‘ž is a polynomial multiple of ๐‘ . Thus there exists a polynomial ๐‘  โˆˆ ๐’ซ (๐…) such that ๐‘ž = ๐‘๐‘  . We have ๐‘ž(๐‘‡) = ๐‘(๐‘‡)๐‘ (๐‘‡) = 0 ๐‘ (๐‘‡) = 0 , as desired. The next result is a nice consequence of the result above. 5.31 minimal polynomial of a restriction operator Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ˆ is a subspace of ๐‘‰ that is invariant under ๐‘‡ . Then the minimal polynomial of ๐‘‡ is a polynomial multiple of the minimal polynomial of ๐‘‡| ๐‘ˆ . Proof Suppose ๐‘ is the minimal polynomial of ๐‘‡ . Thus ๐‘(๐‘‡)๐‘ฃ = 0 for all ๐‘ฃ โˆˆ ๐‘‰ . In particular, ๐‘(๐‘‡)๐‘ข = 0 for all ๐‘ข โˆˆ ๐‘ˆ. Thus ๐‘(๐‘‡| ๐‘ˆ ) = 0 . Now 5.29, applied to the operator ๐‘‡| ๐‘ˆ in place of ๐‘‡ , implies that ๐‘ is a polynomial multiple of the minimal polynomial of ๐‘‡| ๐‘ˆ . See Exercise 25 for a result about quotient operators that is analogous to the result above. The next result shows that the constant term of the minimal polynomial of an operator determines whether the operator is invertible. Section 5B The Minimal Polynomial 149 5.32 ๐‘‡ not invertible โŸบ constant term of minimal polynomial of ๐‘‡ is 0 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then ๐‘‡ is not invertible if and only if the constant term of the minimal polynomial of ๐‘‡ is 0 . Proof Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ is the minimal polynomial of ๐‘‡ . Then ๐‘‡ is not invertible โŸบ 0 is an eigenvalue of ๐‘‡ โŸบ 0 is a zero of ๐‘ โŸบ the constant term of ๐‘ is 0 , where the first equivalence holds by 5.7, the second equivalence holds by 5.27(a), and the last equivalence holds because the constant term of ๐‘ equals ๐‘(0) . Eigenvalues on Odd-Dimensional Real Vector Spaces The next result will be the key tool that we use to show that every operator on an odd-dimensional real vector space has an eigenvalue. 5.33 even-dimensional null space Suppose ๐… = ๐‘ and ๐‘‰ is finite-dimensional. Suppose also that ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ , ๐‘ โˆˆ ๐‘ with ๐‘ 2 < 4๐‘ . Then dim null (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) is an even number. Proof Recall that null (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) is invariant under ๐‘‡ (by 5.18). By replacing ๐‘‰ with null (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) and replacing ๐‘‡ with ๐‘‡ restricted to null (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) , we can assume that ๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ = 0 ; we now need to prove that dim ๐‘‰ is even. Suppose ๐œ† โˆˆ ๐‘ and ๐‘ฃ โˆˆ ๐‘‰ are such that ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ . Then 0 = (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ)๐‘ฃ = (๐œ† 2 + ๐‘๐œ† + ๐‘)๐‘ฃ = ((๐œ† + ๐‘2 ) 2 + ๐‘ โˆ’ ๐‘ 2 4 )๐‘ฃ. The term in large parentheses above is a positive number. Thus the equation above implies that ๐‘ฃ = 0 . Hence we have shown that ๐‘‡ has no eigenvectors. Let ๐‘ˆ be a subspace of ๐‘‰ that is invariant under ๐‘‡ and has the largest dimension among all subspaces of ๐‘‰ that are invariant under ๐‘‡ and have even dimension. If ๐‘ˆ = ๐‘‰ , then we are done; otherwise assume there exists ๐‘ค โˆˆ ๐‘‰ such that ๐‘ค โˆ‰ ๐‘ˆ . Let ๐‘Š = span (๐‘ค , ๐‘‡๐‘ค) . Then ๐‘Š is invariant under ๐‘‡ because ๐‘‡(๐‘‡๐‘ค) = โˆ’๐‘๐‘‡๐‘ค โˆ’ ๐‘๐‘ค . Furthermore, dim ๐‘Š = 2 because otherwise ๐‘ค would be an eigenvector of ๐‘‡ . Now dim (๐‘ˆ + ๐‘Š) = dim ๐‘ˆ + dim ๐‘Š โˆ’ dim (๐‘ˆ โˆฉ ๐‘Š) = dim ๐‘ˆ + 2 , where ๐‘ˆ โˆฉ ๐‘Š = {0} because otherwise ๐‘ˆ โˆฉ ๐‘Š would be a one-dimensional subspace of ๐‘‰ that is invariant under ๐‘‡ (impossible because ๐‘‡ has no eigenvectors). Because ๐‘ˆ + ๐‘Š is invariant under ๐‘‡ , the equation above shows that there exists a subspace of ๐‘‰ invariant under ๐‘‡ of even dimension larger than dim ๐‘ˆ . Thus the assumption that ๐‘ˆ โ‰  ๐‘‰ was incorrect. Hence ๐‘‰ has even dimension. 150 Chapter 5 Eigenvalues and Eigenvectors The next result states that on odd-dimensional vector spaces, every operator has an eigenvalue. We already know this result for finite-dimensional complex vectors spaces (without the odd hypothesis). Thus in the proof below, we will assume that ๐… = ๐‘ . 5.34 operators on odd-dimensional vector spaces have eigenvalues Every operator on an odd-dimensional vector space has an eigenvalue. Proof Suppose ๐… = ๐‘ and ๐‘‰ is finite-dimensional. Let ๐‘› = dim ๐‘‰ , and suppose ๐‘› is an odd number. Let ๐‘‡ โˆˆ โ„’ (๐‘‰) . We will use induction on ๐‘› in steps of size two to show that ๐‘‡ has an eigenvalue. To get started, note that the desired result holds if dim ๐‘‰ = 1 because then every nonzero vector in ๐‘‰ is an eigenvector of ๐‘‡ . Now suppose that ๐‘› โ‰ฅ 3 and the desired result holds for all operators on all odd-dimensional vector spaces of dimension less than ๐‘› . Let ๐‘ denote the minimal polynomial of ๐‘‡ . If ๐‘ is a polynomial multiple of ๐‘ฅ โˆ’ ๐œ† for some ๐œ† โˆˆ ๐‘ , then ๐œ† is an eigenvalue of ๐‘‡ [ by 5.27(a) ] and we are done. Thus we can assume that there exist ๐‘ , ๐‘ โˆˆ ๐‘ such that ๐‘ 2 < 4๐‘ and ๐‘ is a polynomial multiple of ๐‘ฅ 2 + ๐‘๐‘ฅ + ๐‘ (see 4.16). There exists a monic polynomial ๐‘ž โˆˆ ๐’ซ (๐‘) such that ๐‘(๐‘ฅ) = ๐‘ž(๐‘ฅ)(๐‘ฅ 2 + ๐‘๐‘ฅ + ๐‘) for all ๐‘ฅ โˆˆ ๐‘ . Now 0 = ๐‘(๐‘‡) = (๐‘ž(๐‘‡))(๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) , which means that ๐‘ž(๐‘‡) equals 0 on range (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) . Because deg ๐‘ž < deg ๐‘ and ๐‘ is the minimal polynomial of ๐‘‡ , this implies that range (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) โ‰  ๐‘‰ . The fundamental theorem of linear maps (3.21) tells us that dim ๐‘‰ = dim null (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) + dim range (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ). Because dim ๐‘‰ is odd (by hypothesis) and dim null (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) is even (by 5.33), the equation above shows that dim range (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) is odd. Hence range (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) is a subspace of ๐‘‰ that is invariant under ๐‘‡ (by 5.18) and that has odd dimension less than dim ๐‘‰ . Our induction hypothesis now implies that ๐‘‡ restricted to range (๐‘‡ 2 + ๐‘๐‘‡ + ๐‘๐ผ) has an eigenvalue, which means that ๐‘‡ has an eigenvalue. See Exercise 23 in Section 8B and Exercise 10 in Section 9C for alternative proofs of the result above. Exercises 5B 1 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that 9 is an eigenvalue of ๐‘‡ 2 if and only if 3 or โˆ’3 is an eigenvalue of ๐‘‡ . 2 Suppose ๐‘‰ is a complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) has no eigenvalues. Prove that every subspace of ๐‘‰ invariant under ๐‘‡ is either {0} or infinite- dimensional. 3 Suppose ๐‘› is a positive integer and ๐‘‡ โˆˆ โ„’ (๐… ๐‘› ) is defined by ๐‘‡(๐‘ฅ 1 , โ€ฆ , ๐‘ฅ ๐‘› ) = (๐‘ฅ 1 + โ‹ฏ + ๐‘ฅ ๐‘› , โ€ฆ , ๐‘ฅ 1 + โ‹ฏ + ๐‘ฅ ๐‘› ). (a) Find all eigenvalues and eigenvectors of ๐‘‡ . (b) Find the minimal polynomial of ๐‘‡ . The matrix of ๐‘‡ with respect to the standard basis of ๐… ๐‘› consists of all 1 โ€™s. 4 Suppose ๐… = ๐‚ , ๐‘‡ โˆˆ โ„’ (๐‘‰) , ๐‘ โˆˆ ๐’ซ (๐‚) , and ๐›ผ โˆˆ ๐‚ . Prove that ๐›ผ is an eigenvalue of ๐‘(๐‘‡) if and only if ๐›ผ = ๐‘(๐œ†) for some eigenvalue ๐œ† of ๐‘‡ . 5 Give an example of an operator on ๐‘ 2 that shows the result in Exercise 4 does not hold if ๐‚ is replaced with ๐‘ . 6 Suppose ๐‘‡ โˆˆ โ„’ (๐… 2 ) is defined by ๐‘‡(๐‘ค , ๐‘ง) = (โˆ’๐‘ง , ๐‘ค) . Find the minimal polynomial of ๐‘‡ . 7 (a) Give an example of ๐‘† , ๐‘‡ โˆˆ โ„’ (๐… 2 ) such that the minimal polynomial of ๐‘†๐‘‡ does not equal the minimal polynomial of ๐‘‡๐‘† . (b) Suppose ๐‘‰ is finite-dimensional and ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that if at least one of ๐‘† , ๐‘‡ is invertible, then the minimal polynomial of ๐‘†๐‘‡ equals the minimal polynomial of ๐‘‡๐‘† . Hint: Show that if ๐‘† is invertible and ๐‘ โˆˆ ๐’ซ (๐…) , then ๐‘(๐‘‡๐‘†) = ๐‘† โˆ’1 ๐‘(๐‘†๐‘‡)๐‘† . 8 Suppose ๐‘‡ โˆˆ โ„’ (๐‘ 2 ) is the operator of counterclockwise rotation by 1 โˆ˜ . Find the minimal polynomial of ๐‘‡ . Because dim ๐‘ 2 = 2 , the degree of the minimal polynomial of ๐‘‡ is at most 2 . Thus the minimal polynomial of ๐‘‡ is not the tempting polynomial ๐‘ฅ 180 + 1 , even though ๐‘‡ 180 = โˆ’๐ผ . 9 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that with respect to some basis of ๐‘‰ , all entries of the matrix of ๐‘‡ are rational numbers. Explain why all coefficients of the minimal polynomial of ๐‘‡ are rational numbers. 10 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ฃ โˆˆ ๐‘‰ . Prove that span (๐‘ฃ , ๐‘‡๐‘ฃ , โ€ฆ , ๐‘‡ ๐‘š ๐‘ฃ) = span (๐‘ฃ , ๐‘‡๐‘ฃ , โ€ฆ , ๐‘‡ dim ๐‘‰โˆ’1 ๐‘ฃ) for all integers ๐‘š โ‰ฅ dim ๐‘‰ โˆ’ 1 . 11 Suppose ๐‘‰ is a two-dimensional vector space, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and the matrix of ๐‘‡ with respect to some basis of ๐‘‰ is ( ๐‘Ž ๐‘ ๐‘ ๐‘‘ ) . (a) Show that ๐‘‡ 2 โˆ’ (๐‘Ž + ๐‘‘)๐‘‡ + (๐‘Ž๐‘‘ โˆ’ ๐‘๐‘)๐ผ = 0 . (b) Show that the minimal polynomial of ๐‘‡ equals โŽง{ โŽจ{ โŽฉ ๐‘ง โˆ’ ๐‘Ž if ๐‘ = ๐‘ = 0 and ๐‘Ž = ๐‘‘ , ๐‘ง 2 โˆ’ (๐‘Ž + ๐‘‘)๐‘ง + (๐‘Ž๐‘‘ โˆ’ ๐‘๐‘) otherwise . 152 Chapter 5 Eigenvalues and Eigenvectors 12 Define ๐‘‡ โˆˆ โ„’ (๐… ๐‘› ) by ๐‘‡(๐‘ฅ 1 , ๐‘ฅ 2 , ๐‘ฅ 3 , โ€ฆ , ๐‘ฅ ๐‘› ) = (๐‘ฅ 1 , 2๐‘ฅ 2 , 3๐‘ฅ 3 , โ€ฆ , ๐‘›๐‘ฅ ๐‘› ) . Find the minimal polynomial of ๐‘‡ . 13 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ โˆˆ ๐’ซ (๐…) . Prove that there exists a unique ๐‘Ÿ โˆˆ ๐’ซ (๐…) such that ๐‘(๐‘‡) = ๐‘Ÿ(๐‘‡) and deg ๐‘Ÿ is less than the degree of the minimal polynomial of ๐‘‡ . 14 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) has minimal polynomial 4 + 5๐‘ง โˆ’ 6๐‘ง 2 โˆ’ 7๐‘ง 3 + 2๐‘ง 4 + ๐‘ง 5 . Find the minimal polynomial of ๐‘‡ โˆ’1 . 15 Suppose ๐‘‰ is a finite-dimensional complex vector space with dim ๐‘‰ > 0 and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Define ๐‘“ โˆถ ๐‚ โ†’ ๐‘ by ๐‘“ (๐œ†) = dim range (๐‘‡ โˆ’ ๐œ†๐ผ). Prove that ๐‘“ is not a continuous function. 16 Suppose ๐‘Ž 0 , โ€ฆ , ๐‘Ž ๐‘›โˆ’1 โˆˆ ๐… . Let ๐‘‡ be the operator on ๐… ๐‘› whose matrix (with respect to the standard basis) is โŽ›โŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽœโŽ 0 โˆ’๐‘Ž 0 1 0 โˆ’๐‘Ž 1 1 โ‹ฑ โˆ’๐‘Ž 2 โ‹ฑ โ‹ฎ 0 โˆ’๐‘Ž ๐‘›โˆ’2 1 โˆ’๐‘Ž ๐‘›โˆ’1 โŽžโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽ . Here all entries of the matrix are 0 except for all 1 โ€™s on the line under the diagonal and the entries in the last column (some of which might also be 0 ). Show that the minimal polynomial of ๐‘‡ is the polynomial ๐‘Ž 0 + ๐‘Ž 1 ๐‘ง + โ‹ฏ + ๐‘Ž ๐‘›โˆ’1 ๐‘ง ๐‘›โˆ’1 + ๐‘ง ๐‘› . The matrix above is called the companion matrix of the polynomial above. This exercise shows that every monic polynomial is the minimal polynomial of some operator. Hence a formula or an algorithm that could produce exact eigenvalues for each operator on each ๐… ๐‘› could then produce exact zeros for each polynomial [ by 5.27 ( a ) ] . Thus there is no such formula or algorithm. However, efficient numeric methods exist for obtaining very good approximations for the eigenvalues of an operator. 17 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ is the minimal polynomial of ๐‘‡ . Suppose ๐œ† โˆˆ ๐… . Show that the minimal polynomial of ๐‘‡ โˆ’ ๐œ†๐ผ is the polynomial ๐‘ž defined by ๐‘ž(๐‘ง) = ๐‘(๐‘ง + ๐œ†) . 18 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ is the minimal polynomial of ๐‘‡ . Suppose ๐œ† โˆˆ ๐…\{0} . Show that the minimal polynomial of ๐œ†๐‘‡ is the polynomial ๐‘ž defined by ๐‘ž(๐‘ง) = ๐œ† deg ๐‘ ๐‘( ๐‘ง ๐œ† ) . Section 5B The Minimal Polynomial 153 19 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let โ„ฐ be the subspace of โ„’ (๐‘‰) defined by โ„ฐ = {๐‘ž(๐‘‡) โˆถ ๐‘ž โˆˆ ๐’ซ (๐…)}. Prove that dim โ„ฐ equals the degree of the minimal polynomial of ๐‘‡ . 20 Suppose ๐‘‡ โˆˆ โ„’ (๐… 4 ) is such that the eigenvalues of ๐‘‡ are 3 , 5 , 8 . Prove that (๐‘‡ โˆ’ 3๐ผ) 2 (๐‘‡ โˆ’ 5๐ผ) 2 (๐‘‡ โˆ’ 8๐ผ) 2 = 0 . 21 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that the minimal polynomial of ๐‘‡ has degree at most 1 + dim range ๐‘‡ . If dim range ๐‘‡ < dim ๐‘‰ โˆ’ 1 , then this exercise gives a better upper bound than 5.22 for the degree of the minimal polynomial of ๐‘‡ . 22 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ is invertible if and only if ๐ผ โˆˆ span (๐‘‡ , ๐‘‡ 2 , โ€ฆ , ๐‘‡ dim ๐‘‰ ) . 23 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐‘› = dim ๐‘‰ . Prove that if ๐‘ฃ โˆˆ ๐‘‰ , then span (๐‘ฃ , ๐‘‡๐‘ฃ , โ€ฆ , ๐‘‡ ๐‘›โˆ’1 ๐‘ฃ) is invariant under ๐‘‡ . 24 Suppose ๐‘‰ is a finite-dimensional complex vector space. Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is such that 5 and 6 are eigenvalues of ๐‘‡ and that ๐‘‡ has no other eigenvalues. Prove that (๐‘‡ โˆ’ 5๐ผ) dim ๐‘‰โˆ’1 (๐‘‡ โˆ’ 6๐ผ) dim ๐‘‰โˆ’1 = 0 . 25 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ˆ is a subspace of ๐‘‰ that is invariant under ๐‘‡ . (a) Prove that the minimal polynomial of ๐‘‡ is a polynomial multiple of the minimal polynomial of the quotient operator ๐‘‡/๐‘ˆ . (b) Prove that (minimal polynomial of ๐‘‡| ๐‘ˆ ) ร— (minimal polynomial of ๐‘‡/๐‘ˆ ) is a polynomial multiple of the minimal polynomial of ๐‘‡ . The quotient operator ๐‘‡/๐‘ˆ was defined in Exercise 38 in Section 5A. 26 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ˆ is a subspace of ๐‘‰ that is invariant under ๐‘‡ . Prove that the set of eigenvalues of ๐‘‡ equals the union of the set of eigenvalues of ๐‘‡| ๐‘ˆ and the set of eigenvalues of ๐‘‡/๐‘ˆ . 27 Suppose ๐… = ๐‘ , ๐‘‰ is finite-dimensional, and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that the minimal polynomial of ๐‘‡ ๐‚ equals the minimal polynomial of ๐‘‡ . The complexification ๐‘‡ ๐‚ was defined in Exercise 33 of Section 3B. 28 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that the minimal polynomial of ๐‘‡ โ€ฒ โˆˆ โ„’ (๐‘‰ โ€ฒ ) equals the minimal polynomial of ๐‘‡ . The dual map ๐‘‡ โ€ฒ was defined in Section 3F. 29 Show that every operator on a finite-dimensional vector space of dimension at least two has an invariant subspace of dimension two. Exercise 6 in Section 5C will give an improvement of this result when ๐… = ๐‚ . 154 Chapter 5 Eigenvalues and Eigenvectors 5C Upper-Triangular Matrices In Chapter 3 we defined the matrix of a linear map from a finite-dimensional vector space to another finite-dimensional vector space. That matrix depends on a choice of basis of each of the two vector spaces. Now that we are studying operators, which map a vector space to itself, the emphasis is on using only one basis. 5.35 definition: matrix of an operator, โ„ณ (๐‘‡) Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) . The matrix of ๐‘‡ with respect to a basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› of ๐‘‰ is the ๐‘› -by- ๐‘› matrix โ„ณ (๐‘‡) = โŽ›โŽœโŽœโŽœโŽ ๐ด 1 , 1 โ‹ฏ ๐ด 1 , ๐‘› โ‹ฎ โ‹ฎ ๐ด ๐‘› , 1 โ‹ฏ ๐ด ๐‘› , ๐‘› โŽžโŽŸโŽŸโŽŸโŽ  whose entries ๐ด ๐‘— , ๐‘˜ are defined by ๐‘‡๐‘ฃ ๐‘˜ = ๐ด 1 , ๐‘˜ ๐‘ฃ 1 + โ‹ฏ + ๐ด ๐‘› , ๐‘˜ ๐‘ฃ ๐‘› . The notation โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) is used if the basis is not clear from the context. Operators have square matrices (meaning that the number of rows equals the number of columns), rather than the more general rectangular matrices that we considered earlier for linear maps. The ๐‘˜ th column of the matrix โ„ณ (๐‘‡) is formed from the coefficients used to write ๐‘‡๐‘ฃ ๐‘˜ as a linear combination of the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› . If ๐‘‡ is an operator on ๐… ๐‘› and no ba- sis is specified, assume that the basis in question is the standard one ( where the ๐‘˜ th basis vector is 1 in the ๐‘˜ th slot and 0 in all other slots ) . You can then think of the ๐‘˜ th column of โ„ณ (๐‘‡) as ๐‘‡ applied to the ๐‘˜ th basis vector, where we identify ๐‘› -by- 1 column vectors with elements of ๐… ๐‘› . 5.36 example: matrix of an operator with respect to standard basis Define ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘ฅ , ๐‘ฆ , ๐‘ง) = (2๐‘ฅ + ๐‘ฆ , 5๐‘ฆ + 3๐‘ง , 8๐‘ง) . Then the matrix of ๐‘‡ with respect to the standard basis of ๐… 3 is โ„ณ (๐‘‡) = โŽ› โŽœโŽœโŽœโŽ 2 1 0 0 5 3 0 0 8 โŽž โŽŸโŽŸโŽŸโŽ  , as you should verify. A central goal of linear algebra is to show that given an operator ๐‘‡ on a finite- dimensional vector space ๐‘‰ , there exists a basis of ๐‘‰ with respect to which ๐‘‡ has a reasonably simple matrix. To make this vague formulation a bit more precise, we might try to choose a basis of ๐‘‰ such that โ„ณ (๐‘‡) has many 0 โ€™s. Section 5C Upper-Triangular Matrices 155 If ๐‘‰ is a finite-dimensional complex vector space, then we already know enough to show that there is a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ has 0 โ€™s everywhere in the first column, except possibly the first entry. In other words, there is a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ looks like โŽ› โŽœ โŽœโŽœโŽœโŽœโŽœโŽ ๐œ† 0 โˆ— โ‹ฎ 0 โŽž โŽŸ โŽŸโŽŸโŽŸโŽŸโŽŸโŽ  ; here โˆ— denotes the entries in all columns other than the first column. To prove this, let ๐œ† be an eigenvalue of ๐‘‡ (one exists by 5.19) and let ๐‘ฃ be a corresponding eigenvector. Extend ๐‘ฃ to a basis of ๐‘‰ . Then the matrix of ๐‘‡ with respect to this basis has the form above. Soon we will see that we can choose a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ has even more 0 โ€™s. 5.37 definition: diagonal of a matrix The diagonal of a square matrix consists of the entries on the line from the upper left corner to the bottom right corner. For example, the diagonal of the matrix โ„ณ (๐‘‡) = โŽ›โŽœโŽœโŽœโŽ 2 1 0 0 5 3 0 0 8 โŽžโŽŸโŽŸโŽŸโŽ  from Example 5.36 consists of the entries 2 , 5 , 8 , which are shown in red in the matrix above. 5.38 definition: upper-triangular matrix A square matrix is called upper triangular if all entries below the diagonal are 0 . For example, the 3 -by- 3 matrix above is upper triangular. Typically we represent an upper-triangular matrix in the form โŽ› โŽœโŽœโŽœโŽ ๐œ† 1 โˆ— โ‹ฑ 0 ๐œ† ๐‘› โŽž โŽŸโŽŸโŽŸโŽ  ; We often use โˆ— to denote matrix entries that we do not know or that are irrele- vant to the questions being discussed. the 0 in the matrix above indicates that all entries below the diagonal in this ๐‘› -by- ๐‘› matrix equal 0 . Upper-triangular matrices can be considered reasonably simpleโ€”if ๐‘› is large, then at least almost half the entries in an ๐‘› -by- ๐‘› upper- triangular matrix are 0 . 156 Chapter 5 Eigenvalues and Eigenvectors The next result provides a useful connection between upper-triangular matrices and invariant subspaces. 5.39 conditions for upper-triangular matrix Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . Then the following are equivalent. (a) The matrix of ๐‘‡ with respect to ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is upper triangular. (b) span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) is invariant under ๐‘‡ for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . (c) ๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . Proof First suppose (a) holds. To prove that (b) holds, suppose ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘›} . If ๐‘— โˆˆ {1 , โ€ฆ , ๐‘›} , then ๐‘‡๐‘ฃ ๐‘— โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘— ) because the matrix of ๐‘‡ with respect to ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is upper triangular. Because span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘— ) โІ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) if ๐‘— โ‰ค ๐‘˜ , we see that ๐‘‡๐‘ฃ ๐‘— โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) for each ๐‘— โˆˆ {1 , โ€ฆ , ๐‘˜} . Thus span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) is invariant under ๐‘‡ , completing the proof that (a) implies (b). Now suppose (b) holds, so span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) is invariant under ๐‘‡ for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . In particular, ๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . Thus (b) implies (c). Now suppose (c) holds, so ๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . This means that when writing each ๐‘‡๐‘ฃ ๐‘˜ as a linear combination of the basis vectors ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› , we need to use only the vectors ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ . Hence all entries under the diagonal of โ„ณ (๐‘‡) are 0 . Thus โ„ณ (๐‘‡) is an upper-triangular matrix, completing the proof that (c) implies (a). We have shown that (a) โŸน (b) โŸน (c) โŸน (a), which shows that (a), (b), and (c) are equivalent. The next result tells us that if ๐‘‡ โˆˆ โ„’ (๐‘‰) and with respect to some basis of ๐‘‰ we have โ„ณ (๐‘‡) = โŽ›โŽœโŽœโŽœโŽ ๐œ† 1 โˆ— โ‹ฑ 0 ๐œ† ๐‘› โŽžโŽŸโŽŸโŽŸโŽ  , then ๐‘‡ satisfies a simple equation depending on ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› . 5.40 equation satisfied by operator with upper-triangular matrix Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘‰ has a basis with respect to which ๐‘‡ has an upper- triangular matrix with diagonal entries ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› . Then (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘› ๐ผ) = 0. Section 5C Upper-Triangular Matrices 157 Proof Let ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› denote a basis of ๐‘‰ with respect to which ๐‘‡ has an upper- triangular matrix with diagonal entries ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› . Then ๐‘‡๐‘ฃ 1 = ๐œ† 1 ๐‘ฃ 1 , which means that (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)๐‘ฃ 1 = 0 , which implies that (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ 1 = 0 for ๐‘š = 1 , โ€ฆ , ๐‘› (using the commutativity of each ๐‘‡ โˆ’ ๐œ† ๐‘— ๐ผ with each ๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ ). Note that (๐‘‡ โˆ’ ๐œ† 2 ๐ผ)๐‘ฃ 2 โˆˆ span (๐‘ฃ 1 ) . Thus (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)(๐‘‡ โˆ’ ๐œ† 2 ๐ผ)๐‘ฃ 2 = 0 (by the previous paragraph), which implies that (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ 2 = 0 for ๐‘š = 2 , โ€ฆ , ๐‘› (using the commutativity of each ๐‘‡ โˆ’ ๐œ† ๐‘— ๐ผ with each ๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ ). Note that (๐‘‡ โˆ’ ๐œ† 3 ๐ผ)๐‘ฃ 3 โˆˆ span (๐‘ฃ 1 , ๐‘ฃ 2 ) . Thus by the previous paragraph, (๐‘‡โˆ’๐œ† 1 ๐ผ)(๐‘‡โˆ’๐œ† 2 ๐ผ)(๐‘‡โˆ’๐œ† 3 ๐ผ)๐‘ฃ 3 = 0 , which implies that (๐‘‡โˆ’๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡โˆ’๐œ† ๐‘š ๐ผ)๐‘ฃ 3 = 0 for ๐‘š = 3 , โ€ฆ , ๐‘› (using the commutativity of each ๐‘‡ โˆ’ ๐œ† ๐‘— ๐ผ with each ๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ ). Continuing this pattern, we see that (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘› ๐ผ)๐‘ฃ ๐‘˜ = 0 for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . Thus (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘› ๐ผ) is the 0 operator because it is 0 on each vector in a basis of ๐‘‰ . Unfortunately no method exists for exactly computing the eigenvalues of an operator from its matrix. However, if we are fortunate enough to find a basis with respect to which the matrix of the operator is upper triangular, then the problem of computing the eigenvalues becomes trivial, as the next result shows. 5.41 determination of eigenvalues from upper-triangular matrix Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) has an upper-triangular matrix with respect to some basis of ๐‘‰ . Then the eigenvalues of ๐‘‡ are precisely the entries on the diagonal of that upper-triangular matrix. Proof Suppose ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ with respect to which ๐‘‡ has an upper- triangular matrix โ„ณ (๐‘‡) = โŽ›โŽœโŽœโŽœโŽ ๐œ† 1 โˆ— โ‹ฑ 0 ๐œ† ๐‘› โŽžโŽŸโŽŸโŽŸโŽ . Because ๐‘‡๐‘ฃ 1 = ๐œ† 1 ๐‘ฃ 1 , we see that ๐œ† 1 is an eigenvalue of ๐‘‡ . Suppose ๐‘˜ โˆˆ {2 , โ€ฆ , ๐‘›} . Then (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜โˆ’1 ) . Thus ๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ maps span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) into span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜โˆ’1 ) . Because dim span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) = ๐‘˜ and dim span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜โˆ’1 ) = ๐‘˜ โˆ’ 1 , this implies that ๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ restricted to span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) is not injective (by 3.22). Thus there exists ๐‘ฃ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) such that ๐‘ฃ โ‰  0 and (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)๐‘ฃ = 0 . Thus ๐œ† ๐‘˜ is an eigenvalue of ๐‘‡ . Hence we have shown that every entry on the diagonal of โ„ณ (๐‘‡) is an eigenvalue of ๐‘‡ . To prove ๐‘‡ has no other eigenvalues, let ๐‘ž be the polynomial defined by ๐‘ž(๐‘ง) = (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘› ) . Then ๐‘ž(๐‘‡) = 0 (by 5.40). Hence ๐‘ž is a polynomial multiple of the minimal polynomial of ๐‘‡ (by 5.29). Thus every zero of the minimal polynomial of ๐‘‡ is a zero of ๐‘ž . Because the zeros of the minimal polynomial of ๐‘‡ are the eigenvalues of ๐‘‡ (by 5.27), this implies that every eigenvalue of ๐‘‡ is a zero of ๐‘ž . Hence the eigenvalues of ๐‘‡ are all contained in the list ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› . 158 Chapter 5 Eigenvalues and Eigenvectors 5.42 example: eigenvalues via an upper-triangular matrix Define ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘ฅ , ๐‘ฆ , ๐‘ง) = (2๐‘ฅ + ๐‘ฆ , 5๐‘ฆ + 3๐‘ง , 8๐‘ง) . The matrix of ๐‘‡ with respect to the standard basis is โ„ณ (๐‘‡) = โŽ› โŽœโŽœโŽœโŽ 2 1 0 0 5 3 0 0 8 โŽž โŽŸโŽŸโŽŸโŽ  . Now 5.41 implies that the eigenvalues of ๐‘‡ are 2 , 5 , and 8 . The next example illustrates 5.44: an operator has an upper-triangular matrix with respect to some basis if and only if the minimal polynomial of the operator is the product of polynomials of degree 1 . 5.43 example: whether ๐‘‡ has an upper-triangular matrix can depend on ๐… Define ๐‘‡ โˆˆ โ„’ (๐… 4 ) by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 ) = (โˆ’๐‘ง 2 , ๐‘ง 1 , 2๐‘ง 1 + 3๐‘ง 3 , ๐‘ง 3 + 3๐‘ง 4 ). Thus with respect to the standard basis of ๐… 4 , the matrix of ๐‘‡ is โŽ›โŽœโŽœโŽœโŽœโŽœโŽœโŽ 0 โˆ’1 0 0 1 0 0 0 2 0 3 0 0 0 1 3 โŽžโŽŸโŽŸโŽŸโŽŸโŽŸโŽŸโŽ . You can ask a computer to verify that the minimal polynomial of ๐‘‡ is the polyno- mial ๐‘ defined by ๐‘(๐‘ง) = 9 โˆ’ 6๐‘ง + 10๐‘ง 2 โˆ’ 6๐‘ง 3 + ๐‘ง 4 . First consider the case ๐… = ๐‘ . Then the polynomial ๐‘ factors as ๐‘(๐‘ง) = (๐‘ง 2 + 1)(๐‘ง โˆ’ 3)(๐‘ง โˆ’ 3) , with no further factorization of ๐‘ง 2 + 1 as the product of two polynomials of degree 1 with real coefficients. Thus 5.44 states that there does not exist a basis of ๐‘ 4 with respect to which ๐‘‡ has an upper-triangular matrix. Now consider the case ๐… = ๐‚ . Then the polynomial ๐‘ factors as ๐‘(๐‘ง) = (๐‘ง โˆ’ ๐‘–)(๐‘ง + ๐‘–)(๐‘ง โˆ’ 3)(๐‘ง โˆ’ 3) , where all factors above have the form ๐‘งโˆ’๐œ† ๐‘˜ . Thus 5.44 states that there is a basis of ๐‚ 4 with respect to which ๐‘‡ has an upper-triangular matrix. Indeed, you can verify that with respect to the basis (4โˆ’3๐‘– , โˆ’3โˆ’4๐‘– , โˆ’3 + ๐‘– , 1) , (4 + 3๐‘– , โˆ’3 + 4๐‘– , โˆ’3โˆ’๐‘– , 1) , (0 , 0 , 0 , 1) , (0 , 0 , 1 , 0) of ๐‚ 4 , the operator ๐‘‡ has the upper-triangular matrix โŽ›โŽœโŽœ โŽœโŽœโŽœโŽœโŽ ๐‘– 0 0 0 0 โˆ’๐‘– 0 0 0 0 3 1 0 0 0 3 โŽžโŽŸโŽŸ โŽŸโŽŸโŽŸโŽŸโŽ . Section 5C Upper-Triangular Matrices 159 5.44 necessary and sufficient condition to have an upper-triangular matrix Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then ๐‘‡ has an upper- triangular matrix with respect to some basis of ๐‘‰ if and only if the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… . Proof First suppose ๐‘‡ has an upper-triangular matrix with respect to some basis of ๐‘‰ . Let ๐›ผ 1 , โ€ฆ , ๐›ผ ๐‘› denote the diagonal entries of that matrix. Define a polynomial ๐‘ž โˆˆ ๐’ซ (๐…) by ๐‘ž(๐‘ง) = (๐‘ง โˆ’ ๐›ผ 1 )โ‹ฏ(๐‘ง โˆ’ ๐›ผ ๐‘› ). Then ๐‘ž(๐‘‡) = 0 , by 5.40. Hence ๐‘ž is a polynomial multiple of the minimal polyno- mial of ๐‘‡ , by 5.29. Thus the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… with {๐œ† 1 , โ€ฆ , ๐œ† ๐‘š } โІ {๐›ผ 1 , โ€ฆ , ๐›ผ ๐‘› } . To prove the implication in the other direction, now suppose the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… . We will use induction on ๐‘š . To get started, if ๐‘š = 1 then ๐‘ง โˆ’ ๐œ† 1 is the minimal polynomial of ๐‘‡ , which implies that ๐‘‡ = ๐œ† 1 ๐ผ , which implies that the matrix of ๐‘‡ (with respect to any basis of ๐‘‰ ) is upper triangular. Now suppose ๐‘š > 1 and the desired result holds for all smaller positive integers. Let ๐‘ˆ = range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ). Then ๐‘ˆ is invariant under ๐‘‡ [ this is a special case of 5.18 with ๐‘(๐‘ง) = ๐‘ง โˆ’ ๐œ† ๐‘š ] . Thus ๐‘‡| ๐‘ˆ is an operator on ๐‘ˆ . If ๐‘ข โˆˆ ๐‘ˆ , then ๐‘ข = (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ for some ๐‘ฃ โˆˆ ๐‘‰ and (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘šโˆ’1 ๐ผ)๐‘ข = (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ = 0. Hence (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘šโˆ’1 ) is a polynomial multiple of the minimal polynomial of ๐‘‡| ๐‘ˆ , by 5.29. Thus the minimal polynomial of ๐‘‡| ๐‘ˆ is the product of at most ๐‘š โˆ’ 1 terms of the form ๐‘ง โˆ’ ๐œ† ๐‘˜ . By our induction hypothesis, there is a basis ๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘€ of ๐‘ˆ with respect to which ๐‘‡| ๐‘ˆ has an upper-triangular matrix. Thus for each ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘€} , we have (using 5.39) 5.45 ๐‘‡๐‘ข ๐‘˜ = (๐‘‡| ๐‘ˆ )(๐‘ข ๐‘˜ ) โˆˆ span (๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘˜ ). Extend ๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘€ to a basis ๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘€ , ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘ of ๐‘‰ . For each ๐‘˜ โˆˆ {1 , โ€ฆ , ๐‘} , we have ๐‘‡๐‘ฃ ๐‘˜ = (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ ๐‘˜ + ๐œ† ๐‘š ๐‘ฃ ๐‘˜ . The definition of ๐‘ˆ shows that (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ ๐‘˜ โˆˆ ๐‘ˆ = span (๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘€ ) . Thus the equation above shows that 5.46 ๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘€ , ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ). From 5.45 and 5.46, we conclude (using 5.39) that ๐‘‡ has an upper-triangular matrix with respect to the basis ๐‘ข 1 , โ€ฆ , ๐‘ข ๐‘€ , ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘ of ๐‘‰ , as desired. 160 Chapter 5 Eigenvalues and Eigenvectors The set of numbers {๐œ† 1 , โ€ฆ , ๐œ† ๐‘š } from the previous result equals the set of eigenvalues of ๐‘‡ (because the set of zeros of the minimal polynomial of ๐‘‡ equals the set of eigenvalues of ๐‘‡ , by 5.27), although the list ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š in the previous result may contain repetitions. In Chapter 8 we will improve even the wonderful result below; see 8.37 and 8.46. 5.47 if ๐… = ๐‚ , then every operator on ๐‘‰ has an upper-triangular matrix Suppose ๐‘‰ is a finite-dimensional complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then ๐‘‡ has an upper-triangular matrix with respect to some basis of ๐‘‰ . Proof The desired result follows immediately from 5.44 and the second version of the fundamental theorem of algebra (see 4.13). For an extension of the result above to two operators ๐‘† and ๐‘‡ such that ๐‘†๐‘‡ = ๐‘‡๐‘† , see 5.80. Also, for an extension to more than two operators, see Exercise 9(b) in Section 5E. Caution: If an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) has a upper-triangular matrix with respect to some basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› of ๐‘‰ , then the eigenvalues of ๐‘‡ are exactly the entries on the diagonal of โ„ณ (๐‘‡) , as shown by 5.41, and furthermore ๐‘ฃ 1 is an eigenvector of ๐‘‡ . However, ๐‘ฃ 2 , โ€ฆ , ๐‘ฃ ๐‘› need not be eigenvectors of ๐‘‡ . Indeed, a basis vector ๐‘ฃ ๐‘˜ is an eigenvector of ๐‘‡ if and only if all entries in the ๐‘˜ th column of the matrix of ๐‘‡ are 0 , except possibly the ๐‘˜ th entry. The row echelon form of the matrix of an operator does not give us a list of the eigenvalues of the operator. In contrast, an upper-triangular matrix with respect to some basis gives us a list of all the eigenvalues of the op- erator. However, there is no method for computing exactly such an upper- triangular matrix, even though 5.47 guarantees its existence if ๐… = ๐‚ . You may recall from a previous course that every matrix of numbers can be changed to a matrix in what is called row echelon form. If one begins with a square matrix, the matrix in row echelon form will be an upper-triangular matrix. Do not confuse this upper-triangular ma- trix with the upper-triangular matrix of an operator with respect to some basis whose existence is proclaimed by 5.47 (if ๐… = ๐‚ )โ€”there is no connection between these upper-triangular matrices. Exercises 5C 1 Prove or give a counterexample: If ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘‡ 2 has an upper-triangular matrix with respect to some basis of ๐‘‰ , then ๐‘‡ has an upper-triangular matrix with respect to some basis of ๐‘‰ . Section 5C Upper-Triangular Matrices 161 2 Suppose ๐ด and ๐ต are upper-triangular matrices of the same size, with ๐›ผ 1 , โ€ฆ , ๐›ผ ๐‘› on the diagonal of ๐ด and ๐›ฝ 1 , โ€ฆ , ๐›ฝ ๐‘› on the diagonal of ๐ต . (a) Show that ๐ด + ๐ต is an upper-triangular matrix with ๐›ผ 1 + ๐›ฝ 1 , โ€ฆ , ๐›ผ ๐‘› + ๐›ฝ ๐‘› on the diagonal. (b) Show that ๐ด๐ต is an upper-triangular matrix with ๐›ผ 1 ๐›ฝ 1 , โ€ฆ , ๐›ผ ๐‘› ๐›ฝ ๐‘› on the diagonal. The results in this exercise are used in the proof of 5.81. 3 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ with respect to which the matrix of ๐‘‡ is upper triangular, with ๐œ† 1 , โ€ฆ , ๐œ† ๐‘› on the diagonal. Show that the matrix of ๐‘‡ โˆ’1 is also upper triangular with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› , with 1 ๐œ† 1 , โ€ฆ , 1 ๐œ† ๐‘› on the diagonal. 4 Give an example of an operator whose matrix with respect to some basis contains only 0 โ€™s on the diagonal, but the operator is invertible. This exercise and the exercise below show that 5.41 fails without the hypoth- esis that an upper-triangular matrix is under consideration. 5 Give an example of an operator whose matrix with respect to some basis contains only nonzero numbers on the diagonal, but the operator is not invertible. 6 Suppose ๐… = ๐‚ , ๐‘‰ is finite-dimensional, and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that if ๐‘˜ โˆˆ {1 , โ€ฆ , dim ๐‘‰} , then ๐‘‰ has a ๐‘˜ -dimensional subspace invariant under ๐‘‡ . 7 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ฃ โˆˆ ๐‘‰ . (a) Prove that there exists a unique monic polynomial ๐‘ ๐‘ฃ of smallest degree such that ๐‘ ๐‘ฃ (๐‘‡)๐‘ฃ = 0 . (b) Prove that the minimal polynomial of ๐‘‡ is a polynomial multiple of ๐‘ ๐‘ฃ . 8 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and there exists a nonzero vector ๐‘ฃ โˆˆ ๐‘‰ such that ๐‘‡ 2 ๐‘ฃ + 2๐‘‡๐‘ฃ = โˆ’2๐‘ฃ . (a) Prove that if ๐… = ๐‘ , then there does not exist a basis of ๐‘‰ with respect to which ๐‘‡ has an upper-triangular matrix. (b) Prove that if ๐… = ๐‚ and ๐ด is an upper-triangular matrix that equals the matrix of ๐‘‡ with respect to some basis of ๐‘‰ , then โˆ’1 + ๐‘– or โˆ’1 โˆ’ ๐‘– appears on the diagonal of ๐ด . 9 Suppose ๐ต is a square matrix with complex entries. Prove that there exists an invertible square matrix ๐ด with complex entries such that ๐ด โˆ’1 ๐ต๐ด is an upper-triangular matrix. 162 Chapter 5 Eigenvalues and Eigenvectors 10 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . Show that the following are equivalent. (a) The matrix of ๐‘‡ with respect to ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is lower triangular. (b) span (๐‘ฃ ๐‘˜ , โ€ฆ , ๐‘ฃ ๐‘› ) is invariant under ๐‘‡ for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . (c) ๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ ๐‘˜ , โ€ฆ , ๐‘ฃ ๐‘› ) for each ๐‘˜ = 1 , โ€ฆ , ๐‘› . A square matrix is called lower triangular if all entries above the diagonal are 0 . 11 Suppose ๐… = ๐‚ and ๐‘‰ is finite-dimensional. Prove that if ๐‘‡ โˆˆ โ„’ (๐‘‰) , then there exists a basis of ๐‘‰ with respect to which ๐‘‡ has a lower-triangular matrix. 12 Suppose ๐‘‰ is finite-dimensional, ๐‘‡ โˆˆ โ„’ (๐‘‰) has an upper-triangular matrix with respect to some basis of ๐‘‰ , and ๐‘ˆ is a subspace of ๐‘‰ that is invariant under ๐‘‡ . (a) Prove that ๐‘‡| ๐‘ˆ has an upper-triangular matrix with respect to some basis of ๐‘ˆ . (b) Prove that the quotient operator ๐‘‡/๐‘ˆ has an upper-triangular matrix with respect to some basis of ๐‘‰/๐‘ˆ . The quotient operator ๐‘‡/๐‘ˆ was defined in Exercise 38 in Section 5A. 13 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Suppose there exists a subspace ๐‘ˆ of ๐‘‰ that is invariant under ๐‘‡ such that ๐‘‡| ๐‘ˆ has an upper- triangular matrix with respect to some basis of ๐‘ˆ and also ๐‘‡/๐‘ˆ has an upper-triangular matrix with respect to some basis of ๐‘‰/๐‘ˆ . Prove that ๐‘‡ has an upper-triangular matrix with respect to some basis of ๐‘‰ . 14 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ has an upper- triangular matrix with respect to some basis of ๐‘‰ if and only if the dual operator ๐‘‡ โ€ฒ has an upper-triangular matrix with respect to some basis of the dual space ๐‘‰ โ€ฒ . Section 5D Diagonalizable Operators 163 5D Diagonalizable Operators Diagonal Matrices 5.48 definition: diagonal matrix A diagonal matrix is a square matrix that is 0 everywhere except possibly on the diagonal. 5.49 example: diagonal matrix โŽ›โŽœโŽœโŽœโŽ 8 0 0 0 5 0 0 0 5 โŽžโŽŸโŽŸโŽŸโŽ  is a diagonal matrix. Every diagonal matrix is upper tri- angular. Diagonal matrices typically have many more 0 โ€™s than most upper- triangular matrices of the same size. If an operator has a diagonal matrix with respect to some basis, then the en- tries on the diagonal are precisely the eigenvalues of the operator; this follows from 5.41 (or find an easier direct proof for diagonal matrices). 5.50 definition: diagonalizable An operator on ๐‘‰ is called diagonalizable if the operator has a diagonal matrix with respect to some basis of ๐‘‰ . 5.51 example: diagonalization may require a different basis Define ๐‘‡ โˆˆ โ„’ (๐‘ 2 ) by ๐‘‡(๐‘ฅ , ๐‘ฆ) = (41๐‘ฅ + 7๐‘ฆ , โˆ’20๐‘ฅ + 74๐‘ฆ). The matrix of ๐‘‡ with respect to the standard basis of ๐‘ 2 is ( 41 7 โˆ’20 74 ) , which is not a diagonal matrix. However, ๐‘‡ is diagonalizable. Specifically, the matrix of ๐‘‡ with respect to the basis (1 , 4) , (7 , 5) is ( 69 0 0 46 ) because ๐‘‡(1 , 4) = (69 , 276) = 69(1 , 4) and ๐‘‡(7 , 5) = (322 , 230) = 46(7 , 5) . 164 Chapter 5 Eigenvalues and Eigenvectors For ๐œ† โˆˆ ๐… , we will find it convenient to have a name and a notation for the set of vectors that an operator ๐‘‡ maps to ๐œ† times the vector. 5.52 definition: eigenspace, ๐ธ(๐œ† , ๐‘‡) Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… . The eigenspace of ๐‘‡ corresponding to ๐œ† is the subspace ๐ธ(๐œ† , ๐‘‡) of ๐‘‰ defined by ๐ธ(๐œ† , ๐‘‡) = null (๐‘‡ โˆ’ ๐œ†๐ผ) = {๐‘ฃ โˆˆ ๐‘‰ โˆถ ๐‘‡๐‘ฃ = ๐œ†๐‘ฃ}. Hence ๐ธ(๐œ† , ๐‘‡) is the set of all eigenvectors of ๐‘‡ corresponding to ๐œ† , along with the 0 vector. For ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† โˆˆ ๐… , the set ๐ธ(๐œ† , ๐‘‡) is a subspace of ๐‘‰ because the null space of each linear map on ๐‘‰ is a subspace of ๐‘‰ . The definitions imply that ๐œ† is an eigenvalue of ๐‘‡ if and only if ๐ธ(๐œ† , ๐‘‡) โ‰  {0} . 5.53 example: eigenspaces of an operator Suppose the matrix of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) with respect to a basis ๐‘ฃ 1 , ๐‘ฃ 2 , ๐‘ฃ 3 of ๐‘‰ is the matrix in Example 5.49. Then ๐ธ(8 , ๐‘‡) = span (๐‘ฃ 1 ) , ๐ธ(5 , ๐‘‡) = span (๐‘ฃ 2 , ๐‘ฃ 3 ). If ๐œ† is an eigenvalue of an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) , then ๐‘‡ restricted to ๐ธ(๐œ† , ๐‘‡) is just the operator of multiplication by ๐œ† . 5.54 sum of eigenspaces is a direct sum Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are distinct eigenvalues of ๐‘‡ . Then ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + ๐ธ(๐œ† ๐‘š , ๐‘‡) is a direct sum. Furthermore, if ๐‘‰ is finite-dimensional, then dim ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + dim ๐ธ(๐œ† ๐‘š , ๐‘‡) โ‰ค dim ๐‘‰. Proof To show that ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + ๐ธ(๐œ† ๐‘š , ๐‘‡) is a direct sum, suppose ๐‘ฃ 1 + โ‹ฏ + ๐‘ฃ ๐‘š = 0 , where each ๐‘ฃ ๐‘˜ is in ๐ธ(๐œ† ๐‘˜ , ๐‘‡) . Because eigenvectors corresponding to distinct eigenvalues are linearly independent (by 5.11), this implies that each ๐‘ฃ ๐‘˜ equals 0 . Thus ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + ๐ธ(๐œ† ๐‘š , ๐‘‡) is a direct sum (by 1.45), as desired. Now suppose ๐‘‰ is finite-dimensional. Then dim ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + dim ๐ธ(๐œ† ๐‘š , ๐‘‡) = dim (๐ธ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐ธ(๐œ† ๐‘š , ๐‘‡)) โ‰ค dim ๐‘‰ , where the first line follows from 3.94 and the second line follows from 2.37. Section 5D Diagonalizable Operators 165 Conditions for Diagonalizability The following characterizations of diagonalizable operators will be useful. 5.55 conditions equivalent to diagonalizability Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š denote the distinct eigenvalues of ๐‘‡ . Then the following are equivalent. (a) ๐‘‡ is diagonalizable. (b) ๐‘‰ has a basis consisting of eigenvectors of ๐‘‡ . (c) ๐‘‰ = ๐ธ(๐œ† 1 , ๐‘‡) โŠ• โ‹ฏ โŠ• ๐ธ(๐œ† ๐‘š , ๐‘‡) . (d) dim ๐‘‰ = dim ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + dim ๐ธ(๐œ† ๐‘š , ๐‘‡) . Proof An operator ๐‘‡ โˆˆ โ„’ (๐‘‰) has a diagonal matrix โŽ›โŽœโŽœโŽœโŽ ๐œ† 1 0 โ‹ฑ 0 ๐œ† ๐‘› โŽžโŽŸโŽŸโŽŸโŽ  with respect to a basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› of ๐‘‰ if and only if ๐‘‡๐‘ฃ ๐‘˜ = ๐œ† ๐‘˜ ๐‘ฃ ๐‘˜ for each ๐‘˜ . Thus (a) and (b) are equivalent. Suppose (b) holds; thus ๐‘‰ has a basis consisting of eigenvectors of ๐‘‡ . Hence every vector in ๐‘‰ is a linear combination of eigenvectors of ๐‘‡ , which implies that ๐‘‰ = ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + ๐ธ(๐œ† ๐‘š , ๐‘‡). Now 5.54 shows that (c) holds, proving that (b) implies (c). That (c) implies (d) follows immediately from 3.94. Finally, suppose (d) holds; thus 5.56 dim ๐‘‰ = dim ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + dim ๐ธ(๐œ† ๐‘š , ๐‘‡). Choose a basis of each ๐ธ(๐œ† ๐‘˜ , ๐‘‡) ; put all these bases together to form a list ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› of eigenvectors of ๐‘‡ , where ๐‘› = dim ๐‘‰ (by 5.56). To show that this list is linearly independent, suppose ๐‘Ž 1 ๐‘ฃ 1 + โ‹ฏ + ๐‘Ž ๐‘› ๐‘ฃ ๐‘› = 0 , where ๐‘Ž 1 , โ€ฆ , ๐‘Ž ๐‘› โˆˆ ๐… . For each ๐‘˜ = 1 , โ€ฆ , ๐‘š , let ๐‘ข ๐‘˜ denote the sum of all the terms ๐‘Ž ๐‘— ๐‘ฃ ๐‘— such that ๐‘ฃ ๐‘— โˆˆ ๐ธ(๐œ† ๐‘˜ , ๐‘‡) . Thus each ๐‘ข ๐‘˜ is in ๐ธ(๐œ† ๐‘˜ , ๐‘‡) , and ๐‘ข 1 + โ‹ฏ + ๐‘ข ๐‘š = 0. Because eigenvectors corresponding to distinct eigenvalues are linearly indepen- dent (see 5.11), this implies that each ๐‘ข ๐‘˜ equals 0 . Because each ๐‘ข ๐‘˜ is a sum of terms ๐‘Ž ๐‘— ๐‘ฃ ๐‘— , where the ๐‘ฃ ๐‘— โ€™s were chosen to be a basis of ๐ธ(๐œ† ๐‘˜ , ๐‘‡) , this implies that all ๐‘Ž ๐‘— โ€™s equal 0 . Thus ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is linearly independent and hence is a basis of ๐‘‰ (by 2.38). Thus (d) implies (b), completing the proof. For additional conditions equivalent to diagonalizability, see 5.62, Exercises 5 and 15 in this section, Exercise 24 in Section 7B, and Exercise 15 in Section 8A. 166 Chapter 5 Eigenvalues and Eigenvectors As we know, every operator on a finite-dimensional complex vector space has an eigenvalue. However, not every operator on a finite-dimensional complex vector space has enough eigenvectors to be diagonalizable, as shown by the next example. 5.57 example: an operator that is not diagonalizable Define an operator ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘Ž , ๐‘ , ๐‘) = (๐‘ , ๐‘ , 0) . The matrix of ๐‘‡ with respect to the standard basis of ๐… 3 is โŽ›โŽœโŽœโŽœโŽ 0 1 0 0 0 1 0 0 0 โŽžโŽŸโŽŸโŽŸโŽ  , which is an upper-triangular matrix but is not a diagonal matrix. As you should verify, 0 is the only eigenvalue of ๐‘‡ and furthermore ๐ธ(0 , ๐‘‡) = {(๐‘Ž , 0 , 0) โˆˆ ๐… 3 โˆถ ๐‘Ž โˆˆ ๐…}. Hence conditions (b), (c), and (d) of 5.55 fail (of course, because these conditions are equivalent, it is sufficient to check that only one of them fails). Thus condition (a) of 5.55 also fails. Hence ๐‘‡ is not diagonalizable, regardless of whether ๐… = ๐‘ or ๐… = ๐‚ . The next result shows that if an operator has as many distinct eigenvalues as the dimension of its domain, then the operator is diagonalizable. 5.58 enough eigenvalues implies diagonalizability Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) has dim ๐‘‰ distinct eigenvalues. Then ๐‘‡ is diagonalizable. Proof Suppose ๐‘‡ has distinct eigenvalues ๐œ† 1 , โ€ฆ , ๐œ† dim ๐‘‰ . For each ๐‘˜ , let ๐‘ฃ ๐‘˜ โˆˆ ๐‘‰ be an eigenvector corresponding to the eigenvalue ๐œ† ๐‘˜ . Because eigenvectors corre- sponding to distinct eigenvalues are linearly independent (see 5.11), ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ dim ๐‘‰ is linearly independent. A linearly independent list of dim ๐‘‰ vectors in ๐‘‰ is a basis of ๐‘‰ (see 2.38); thus ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ dim ๐‘‰ is a basis of ๐‘‰ . With respect to this basis consisting of eigenvectors, ๐‘‡ has a diagonal matrix. In later chapters we will find additional conditions that imply that certain operators are diagonalizable. For example, see the real spectral theorem (7.29) and the complex spectral theorem (7.31). The result above gives a sufficient condition for an operator to be diagonal- izable. However, this condition is not necessary. For example, the operator ๐‘‡ on ๐… 3 defined by ๐‘‡(๐‘ฅ , ๐‘ฆ , ๐‘ง) = (6๐‘ฅ , 6๐‘ฆ , 7๐‘ง) has only two eigenvalues ( 6 and 7 ) and dim ๐… 3 = 3 , but ๐‘‡ is diagonalizable ( by the standard basis of ๐… 3 ) . Section 5D Diagonalizable Operators 167 For a spectacular application of these techniques, see Exercise 21, which shows how to use diagonalization to find an exact formula for the ๐‘› th term of the Fibonacci sequence. The next example illustrates the im- portance of diagonalization, which can be used to compute high powers of an operator, taking advantage of the equa- tion ๐‘‡ ๐‘˜ ๐‘ฃ = ๐œ† ๐‘˜ ๐‘ฃ if ๐‘ฃ is an eigenvector of ๐‘‡ with eigenvalue ๐œ† . 5.59 example: using diagonalization to compute ๐‘‡ 100 Define ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘ฅ , ๐‘ฆ , ๐‘ง) = (2๐‘ฅ + ๐‘ฆ , 5๐‘ฆ + 3๐‘ง , 8๐‘ง) . With respect to the standard basis, the matrix of ๐‘‡ is โŽ›โŽœโŽœโŽœโŽ 2 1 0 0 5 3 0 0 8 โŽžโŽŸโŽŸโŽŸโŽ  . The matrix above is an upper-triangular matrix but it is not a diagonal matrix. By 5.41, the eigenvalues of ๐‘‡ are 2 , 5 , and 8 . Because ๐‘‡ is an operator on a vector space of dimension three and ๐‘‡ has three distinct eigenvalues, 5.58 assures us that there exists a basis of ๐… 3 with respect to which ๐‘‡ has a diagonal matrix. To find this basis, we only have to find an eigenvector for each eigenvalue. In other words, we have to find a nonzero solution to the equation ๐‘‡(๐‘ฅ , ๐‘ฆ , ๐‘ง) = ๐œ†(๐‘ฅ , ๐‘ฆ , ๐‘ง) for ๐œ† = 2 , then for ๐œ† = 5 , and then for ๐œ† = 8 . Solving these simple equations shows that for ๐œ† = 2 we have an eigenvector (1 , 0 , 0) , for ๐œ† = 5 we have an eigenvector (1 , 3 , 0) , and for ๐œ† = 8 we have an eigenvector (1 , 6 , 6) . Thus (1 , 0 , 0) , (1 , 3 , 0) , (1 , 6 , 6) is a basis of ๐… 3 consisting of eigenvectors of ๐‘‡ , and with respect to this basis the matrix of ๐‘‡ is the diagonal matrix โŽ›โŽœโŽœโŽœโŽ 2 0 0 0 5 0 0 0 8 โŽžโŽŸโŽŸโŽŸโŽ  . To compute ๐‘‡ 100 (0 , 0 , 1) , for example, write (0 , 0 , 1) as a linear combination of our basis of eigenvectors: (0 , 0 , 1) = 16 (1 , 0 , 0) โˆ’ 13 (1 , 3 , 0) + 16 (1 , 6 , 6). Now apply ๐‘‡ 100 to both sides of the equation above, getting ๐‘‡ 100 (0 , 0 , 1) = 16 (๐‘‡ 100 (1 , 0 , 0)) โˆ’ 13 (๐‘‡ 100 (1 , 3 , 0)) + 16 (๐‘‡ 100 (1 , 6 , 6)) = 16 (2 100 (1 , 0 , 0) โˆ’ 2 โ‹… 5 100 (1 , 3 , 0) + 8 100 (1 , 6 , 6)) = 16 (2 100 โˆ’ 2 โ‹… 5 100 + 8 100 , 6 โ‹… 8 100 โˆ’ 6 โ‹… 5 100 , 6 โ‹… 8 100 ). 168 Chapter 5 Eigenvalues and Eigenvectors We saw earlier that an operator ๐‘‡ on a finite-dimensional vector space ๐‘‰ has an upper-triangular matrix with respect to some basis of ๐‘‰ if and only if the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… (see 5.44). As we previously noted (see 5.47), this condition is always satisfied if ๐… = ๐‚ . Our next result 5.62 states that an operator ๐‘‡ โˆˆ โ„’ (๐‘‰) has a diagonal matrix with respect to some basis of ๐‘‰ if and only if the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some distinct ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… . Before formally stating this result, we give two examples of using it. 5.60 example: diagonalizable, but with no known exact eigenvalues Define ๐‘‡ โˆˆ โ„’ (๐‚ 5 ) by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 , ๐‘ง 5 ) = (โˆ’3๐‘ง 5 , ๐‘ง 1 + 6๐‘ง 5 , ๐‘ง 2 , ๐‘ง 3 , ๐‘ง 4 ). The matrix of ๐‘‡ is shown in Example 5.26, where we showed that the minimal polynomial of ๐‘‡ is 3 โˆ’ 6๐‘ง + ๐‘ง 5 . As mentioned in Example 5.28, no exact expression is known for any of the zeros of this polynomial, but numeric techniques show that the zeros of this polynomial are approximately โˆ’1.67 , 0.51 , 1.40 , โˆ’0.12 + 1.59๐‘– , โˆ’0.12 โˆ’ 1.59๐‘– . The software that produces these approximations is accurate to more than three digits. Thus these approximations are good enough to show that the five numbers above are distinct. The minimal polynomial of ๐‘‡ equals the fifth degree monic polynomial with these zeros. Now 5.62 shows that ๐‘‡ is diagonalizable. 5.61 example: showing that an operator is not diagonalizable Define ๐‘‡ โˆˆ โ„’ (๐… 3 ) by ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (6๐‘ง 1 + 3๐‘ง 2 + 4๐‘ง 3 , 6๐‘ง 2 + 2๐‘ง 3 , 7๐‘ง 3 ). The matrix of ๐‘‡ with respect to the standard basis of ๐… 3 is โŽ›โŽœโŽœโŽœโŽ 6 3 4 0 6 2 0 0 7 โŽžโŽŸโŽŸโŽŸโŽ  . The matrix above is an upper-triangular matrix but is not a diagonal matrix. Might ๐‘‡ have a diagonal matrix with respect to some other basis of ๐… 3 ? To answer this question, we will find the minimal polynomial of ๐‘‡ . First note that the eigenvalues of ๐‘‡ are the diagonal entries of the matrix above (by 5.41). Thus the zeros of the minimal polynomial of ๐‘‡ are 6 , 7 [ by 5.27(a) ] . The diagonal of the matrix above tells us that (๐‘‡ โˆ’ 6๐ผ) 2 (๐‘‡ โˆ’ 7๐ผ) = 0 (by 5.40). The minimal polynomial of ๐‘‡ has degree at most 3 (by 5.22). Putting all this together, we see that the minimal polynomial of ๐‘‡ is either (๐‘ง โˆ’ 6)(๐‘ง โˆ’ 7) or (๐‘ง โˆ’ 6) 2 (๐‘ง โˆ’ 7) . A simple computation shows that (๐‘‡ โˆ’ 6๐ผ)(๐‘‡ โˆ’ 7๐ผ) โ‰  0 . Thus the minimal polynomial of ๐‘‡ is (๐‘ง โˆ’ 6) 2 (๐‘ง โˆ’ 7) . Now 5.62 shows that ๐‘‡ is not diagonalizable. Section 5D Diagonalizable Operators 169 5.62 necessary and sufficient condition for diagonalizability Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Then ๐‘‡ is diagonalizable if and only if the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some list of distinct numbers ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… . Proof First suppose ๐‘‡ is diagonalizable. Thus there is a basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› of ๐‘‰ consisting of eigenvectors of ๐‘‡ . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š be the distinct eigenvalues of ๐‘‡ . Then for each ๐‘ฃ ๐‘— , there exists ๐œ† ๐‘˜ with (๐‘‡ โˆ’ ๐œ† ๐‘˜ ๐ผ)๐‘ฃ ๐‘— = 0 . Thus (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ ๐‘— = 0 , which implies that the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) . To prove the implication in the other direction, now suppose the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some list of distinct numbers ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… . Thus 5.63 (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) = 0. We will prove that ๐‘‡ is diagonalizable by induction on ๐‘š . To get started, suppose ๐‘š = 1 . Then ๐‘‡ โˆ’ ๐œ† 1 ๐ผ = 0 , which means that ๐‘‡ is a scalar multiple of the identity operator, which implies that ๐‘‡ is diagonalizable. Now suppose that ๐‘š > 1 and the desired result holds for all smaller values of ๐‘š . The subspace range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) is invariant under ๐‘‡ [ this is a special case of 5.18 with ๐‘(๐‘ง) = ๐‘ง โˆ’ ๐œ† ๐‘š ] . Thus ๐‘‡ restricted to range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) is an operator on range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) . If ๐‘ข โˆˆ range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) , then ๐‘ข = (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ for some ๐‘ฃ โˆˆ ๐‘‰ , and 5.63 implies 5.64 (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘šโˆ’1 ๐ผ)๐‘ข = (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ)๐‘ฃ = 0. Hence (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘šโˆ’1 ) is a polynomial multiple of the minimal polynomial of ๐‘‡ restricted to range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) [by 5.29]. Thus by our induction hypothesis, there is a basis of range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) consisting of eigenvectors of ๐‘‡ . Suppose that ๐‘ข โˆˆ range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) โˆฉ null (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) . Then ๐‘‡๐‘ข = ๐œ† ๐‘š ๐‘ข . Now 5.64 implies that 0 = (๐‘‡ โˆ’ ๐œ† 1 ๐ผ)โ‹ฏ(๐‘‡ โˆ’ ๐œ† ๐‘šโˆ’1 ๐ผ)๐‘ข = (๐œ† ๐‘š โˆ’ ๐œ† 1 )โ‹ฏ(๐œ† ๐‘š โˆ’ ๐œ† ๐‘šโˆ’1 )๐‘ข. Because ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š are distinct, the equation above implies that ๐‘ข = 0 . Hence range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) โˆฉ null (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) = {0} . Thus range (๐‘‡โˆ’๐œ† ๐‘š ๐ผ) + null (๐‘‡โˆ’๐œ† ๐‘š ๐ผ) is a direct sum (by 1.46) whose dimension is dim ๐‘‰ (by 3.94 and 3.21). Hence range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) โŠ• null (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) = ๐‘‰ . Every vector in null (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) is an eigenvector of ๐‘‡ with eigenvalue ๐œ† ๐‘š . Earlier in this proof we saw that there is a basis of range (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) consisting of eigenvectors of ๐‘‡ . Adjoining to that basis a basis of null (๐‘‡ โˆ’ ๐œ† ๐‘š ๐ผ) gives a basis of ๐‘‰ consisting of eigenvectors of ๐‘‡ . The matrix of ๐‘‡ with respect to this basis is a diagonal matrix, as desired. 170 Chapter 5 Eigenvalues and Eigenvectors No formula exists for the zeros of polynomials of degree 5 or greater. However, the previous result can be used to determine whether an operator on a complex vector space is diagonalizable without even finding approximations of the zeros of the minimal polynomialโ€”see Exercise 15. The next result will be a key tool when we prove a result about the simultaneous diagonalization of two operators; see 5.76. Note how the use of a characterization of diagonalizable operators in terms of the minimal polynomial (see 5.62) leads to a short proof of the next result. 5.65 restriction of diagonalizable operator to invariant subspace Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is diagonalizable and ๐‘ˆ is a subspace of ๐‘‰ that is invariant under ๐‘‡ . Then ๐‘‡| ๐‘ˆ is a diagonalizable operator on ๐‘ˆ . Proof Because the operator ๐‘‡ is diagonalizable, the minimal polynomial of ๐‘‡ equals (๐‘ง โˆ’ ๐œ† 1 )โ‹ฏ(๐‘ง โˆ’ ๐œ† ๐‘š ) for some list of distinct numbers ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š โˆˆ ๐… (by 5.62). The minimal polynomial of ๐‘‡ is a polynomial multiple of the minimal polynomial of ๐‘‡| ๐‘ˆ (by 5.31). Hence the minimal polynomial of ๐‘‡| ๐‘ˆ has the form required by 5.62, which shows that ๐‘‡| ๐‘ˆ is diagonalizable. Gershgorin Disk Theorem 5.66 definition: Gershgorin disks Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . Let ๐ด denote the matrix of ๐‘‡ with respect to this basis. A Gershgorin disk of ๐‘‡ with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a set of the form {๐‘ง โˆˆ ๐… โˆถ |๐‘ง โˆ’ ๐ด ๐‘— , ๐‘— | โ‰ค ๐‘› โˆ‘ ๐‘˜=1๐‘˜โ‰ ๐‘— |๐ด ๐‘— , ๐‘˜ |} , where ๐‘— โˆˆ {1 , โ€ฆ , ๐‘›} . Because there are ๐‘› choices for ๐‘— in the definition above, ๐‘‡ has ๐‘› Gershgorin disks. If ๐… = ๐‚ , then for each ๐‘— โˆˆ {1 , โ€ฆ , ๐‘›} , the corresponding Gershgorin disk is a closed disk in ๐‚ centered at ๐ด ๐‘— , ๐‘— , which is the ๐‘— th entry on the diagonal of ๐ด . The radius of this closed disk is the sum of the absolute values of the entries in row ๐‘— of ๐ด , excluding the diagonal entry. If ๐… = ๐‘ , then the Gershgorin disks are closed intervals in ๐‘ . In the special case that the square matrix ๐ด above is a diagonal matrix, each Gershgorin disk consists of a single point that is a diagonal entry of ๐ด (and each eigenvalue of ๐‘‡ is one of those points, as required by the next result). One consequence of our next result is that if the nondiagonal entries of ๐ด are small, then each eigenvalue of ๐‘‡ is near a diagonal entry of ๐ด . Section 5D Diagonalizable Operators 171 5.67 Gershgorin disk theorem Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . Then each eigenvalue of ๐‘‡ is contained in some Gershgorin disk of ๐‘‡ with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› . Proof Suppose ๐œ† โˆˆ ๐… is an eigenvalue of ๐‘‡ . Let ๐‘ค โˆˆ ๐‘‰ be a corresponding eigenvector. There exist ๐‘ 1 , โ€ฆ , ๐‘ ๐‘› โˆˆ ๐… such that 5.68 ๐‘ค = ๐‘ 1 ๐‘ฃ 1 + โ‹ฏ + ๐‘ ๐‘› ๐‘ฃ ๐‘› . Let ๐ด denote the matrix of ๐‘‡ with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› . Applying ๐‘‡ to both sides of the equation above gives ๐œ†๐‘ค = ๐‘› โˆ‘ ๐‘˜=1 ๐‘ ๐‘˜ ๐‘‡๐‘ฃ ๐‘˜ 5.69 = ๐‘› โˆ‘ ๐‘˜=1 ๐‘ ๐‘˜ ๐‘› โˆ‘ ๐‘— = 1 ๐ด ๐‘— , ๐‘˜ ๐‘ฃ ๐‘— = ๐‘› โˆ‘ ๐‘— = 1 ( ๐‘› โˆ‘ ๐‘˜=1 ๐ด ๐‘— , ๐‘˜ ๐‘ ๐‘˜ )๐‘ฃ ๐‘— . 5.70 Let ๐‘— โˆˆ {1 , โ€ฆ , ๐‘›} be such that |๐‘ ๐‘— | = max {|๐‘ 1 | , โ€ฆ , |๐‘ ๐‘› |}. Using 5.68, we see that the coefficient of ๐‘ฃ ๐‘— on the left side of 5.69 equals ๐œ†๐‘ ๐‘— , which must equal the coefficient of ๐‘ฃ ๐‘— on the right side of 5.70. In other words, ๐œ†๐‘ ๐‘— = ๐‘› โˆ‘ ๐‘˜=1 ๐ด ๐‘— , ๐‘˜ ๐‘ ๐‘˜ . Subtract ๐ด ๐‘— , ๐‘— ๐‘ ๐‘— from each side of the equation above and then divide both sides by ๐‘ ๐‘— to get |๐œ† โˆ’ ๐ด ๐‘— , ๐‘— | = โˆฃ ๐‘› โˆ‘ ๐‘˜=1๐‘˜โ‰ ๐‘— ๐ด ๐‘— , ๐‘˜ ๐‘ ๐‘˜ ๐‘ ๐‘— โˆฃ โ‰ค ๐‘› โˆ‘ ๐‘˜=1๐‘˜โ‰ ๐‘— |๐ด ๐‘— , ๐‘˜ |. Thus ๐œ† is in the ๐‘— th Gershgorin disk with respect to the basis ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› . The Gershgorin disk theorem is named for Semyon Aronovich Gershgorin, who published this result in 1931. Exercise 22 gives a nice application of the Gershgorin disk theorem. Exercise 23 states that the radius of each Gershgorin disk could be changed to the sum of the absolute values of corresponding column entries (instead of row entries), excluding the diagonal entry, and the theorem above would still hold. 172 Chapter 5 Eigenvalues and Eigenvectors Exercises 5D 1 Suppose ๐‘‰ is a finite-dimensional complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) . (a) Prove that if ๐‘‡ 4 = ๐ผ , then ๐‘‡ is diagonalizable. (b) Prove that if ๐‘‡ 4 = ๐‘‡ , then ๐‘‡ is diagonalizable. (c) Give an example of an operator ๐‘‡ โˆˆ โ„’ (๐‚ 2 ) such that ๐‘‡ 4 = ๐‘‡ 2 and ๐‘‡ is not diagonalizable. 2 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) has a diagonal matrix ๐ด with respect to some basis of ๐‘‰ . Prove that if ๐œ† โˆˆ ๐… , then ๐œ† appears on the diagonal of ๐ด precisely dim ๐ธ(๐œ† , ๐‘‡) times. 3 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that if the operator ๐‘‡ is diagonalizable, then ๐‘‰ = null ๐‘‡ โŠ• range ๐‘‡ . 4 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that the following are equivalent. (a) ๐‘‰ = null ๐‘‡ โŠ• range ๐‘‡ . (b) ๐‘‰ = null ๐‘‡ + range ๐‘‡ . (c) null ๐‘‡ โˆฉ range ๐‘‡ = {0} . 5 Suppose ๐‘‰ is a finite-dimensional complex vector space and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ is diagonalizable if and only if ๐‘‰ = null (๐‘‡ โˆ’ ๐œ†๐ผ) โŠ• range (๐‘‡ โˆ’ ๐œ†๐ผ) for every ๐œ† โˆˆ ๐‚ . 6 Suppose ๐‘‡ โˆˆ โ„’ (๐… 5 ) and dim ๐ธ(8 , ๐‘‡) = 4 . Prove that ๐‘‡ โˆ’ 2๐ผ or ๐‘‡ โˆ’ 6๐ผ is invertible. 7 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible. Prove that ๐ธ(๐œ† , ๐‘‡) = ๐ธ( 1๐œ† , ๐‘‡ โˆ’1 ) for every ๐œ† โˆˆ ๐… with ๐œ† โ‰  0 . 8 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š denote the distinct nonzero eigenvalues of ๐‘‡ . Prove that dim ๐ธ(๐œ† 1 , ๐‘‡) + โ‹ฏ + dim ๐ธ(๐œ† ๐‘š , ๐‘‡) โ‰ค dim range ๐‘‡. 9 Suppose ๐‘… , ๐‘‡ โˆˆ โ„’ (๐… 3 ) each have 2 , 6 , 7 as eigenvalues. Prove that there exists an invertible operator ๐‘† โˆˆ โ„’ (๐… 3 ) such that ๐‘… = ๐‘† โˆ’1 ๐‘‡๐‘† . 10 Find ๐‘… , ๐‘‡ โˆˆ โ„’ (๐… 4 ) such that ๐‘… and ๐‘‡ each have 2 , 6 , 7 as eigenvalues, ๐‘… and ๐‘‡ have no other eigenvalues, and there does not exist an invertible operator ๐‘† โˆˆ โ„’ (๐… 4 ) such that ๐‘… = ๐‘† โˆ’1 ๐‘‡๐‘† . Section 5D Diagonalizable Operators 173 11 Find ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) such that 6 and 7 are eigenvalues of ๐‘‡ and such that ๐‘‡ does not have a diagonal matrix with respect to any basis of ๐‚ 3 . 12 Suppose ๐‘‡ โˆˆ โ„’ (๐‚ 3 ) is such that 6 and 7 are eigenvalues of ๐‘‡ . Furthermore, suppose ๐‘‡ does not have a diagonal matrix with respect to any basis of ๐‚ 3 . Prove that there exists (๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) โˆˆ ๐‚ 3 such that ๐‘‡(๐‘ง 1 , ๐‘ง 2 , ๐‘ง 3 ) = (6 + 8๐‘ง 1 , 7 + 8๐‘ง 2 , 13 + 8๐‘ง 3 ). 13 Suppose ๐ด is a diagonal matrix with distinct entries on the diagonal and ๐ต is a matrix of the same size as ๐ด . Show that ๐ด๐ต = ๐ต๐ด if and only if ๐ต is a diagonal matrix. 14 (a) Give an example of a finite-dimensional complex vector space and an operator ๐‘‡ on that vector space such that ๐‘‡ 2 is diagonalizable but ๐‘‡ is not diagonalizable. (b) Suppose ๐… = ๐‚ , ๐‘˜ is a positive integer, and ๐‘‡ โˆˆ โ„’ (๐‘‰) is invertible. Prove that ๐‘‡ is diagonalizable if and only if ๐‘‡ ๐‘˜ is diagonalizable. 15 Suppose ๐‘‰ is a finite-dimensional complex vector space, ๐‘‡ โˆˆ โ„’ (๐‘‰) , and ๐‘ is the minimal polynomial of ๐‘‡ . Prove that the following are equivalent. (a) ๐‘‡ is diagonalizable. (b) There does not exist ๐œ† โˆˆ ๐‚ such that ๐‘ is a polynomial multiple of (๐‘ง โˆ’ ๐œ†) 2 . (c) ๐‘ and its derivative ๐‘ โ€ฒ have no zeros in common. (d) The greatest common divisor of ๐‘ and ๐‘ โ€ฒ is the constant polynomial 1 . The greatest common divisor of ๐‘ and ๐‘ โ€ฒ is the monic polynomial ๐‘ž of largest degree such that ๐‘ and ๐‘ โ€ฒ are both polynomial multiples of ๐‘ž . The Euclidean algorithm for polynomials ( look it up ) can quickly determine the greatest common divisor of two polynomials, without requiring any information about the zeros of the polynomials. Thus the equivalence of ( a ) and ( d ) above shows that we can determine whether ๐‘‡ is diagonalizable without knowing anything about the zeros of ๐‘ . 16 Suppose that ๐‘‡ โˆˆ โ„’ (๐‘‰) is diagonalizable. Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š denote the distinct eigenvalues of ๐‘‡ . Prove that a subspace ๐‘ˆ of ๐‘‰ is invariant under ๐‘‡ if and only if there exist subspaces ๐‘ˆ 1 , โ€ฆ , ๐‘ˆ ๐‘š of ๐‘‰ such that ๐‘ˆ ๐‘˜ โІ ๐ธ(๐œ† ๐‘˜ , ๐‘‡) for each ๐‘˜ and ๐‘ˆ = ๐‘ˆ 1 โŠ• โ‹ฏ โŠ• ๐‘ˆ ๐‘š . 17 Suppose ๐‘‰ is finite-dimensional. Prove that โ„’ (๐‘‰) has a basis consisting of diagonalizable operators. 18 Suppose that ๐‘‡ โˆˆ โ„’ (๐‘‰) is diagonalizable and ๐‘ˆ is a subspace of ๐‘‰ that is invariant under ๐‘‡ . Prove that the quotient operator ๐‘‡/๐‘ˆ is a diagonalizable operator on ๐‘‰/๐‘ˆ . The quotient operator ๐‘‡/๐‘ˆ was defined in Exercise 38 in Section 5A. 174 Chapter 5 Eigenvalues and Eigenvectors 19 Prove or give a counterexample: If ๐‘‡ โˆˆ โ„’ (๐‘‰) and there exists a subspace ๐‘ˆ of ๐‘‰ that is invariant under ๐‘‡ such that ๐‘‡| ๐‘ˆ and ๐‘‡/๐‘ˆ are both diagonalizable, then ๐‘‡ is diagonalizable. See Exercise 13 in Section 5C for an analogous statement about upper- triangular matrices. 20 Suppose ๐‘‰ is finite-dimensional and ๐‘‡ โˆˆ โ„’ (๐‘‰) . Prove that ๐‘‡ is diagonaliz- able if and only if the dual operator ๐‘‡ โ€ฒ is diagonalizable. 21 The Fibonacci sequence ๐น 0 , ๐น 1 , ๐น 2 , โ€ฆ is defined by ๐น 0 = 0 , ๐น 1 = 1 , and ๐น ๐‘› = ๐น ๐‘›โˆ’2 + ๐น ๐‘›โˆ’1 for ๐‘› โ‰ฅ 2. Define ๐‘‡ โˆˆ โ„’ (๐‘ 2 ) by ๐‘‡(๐‘ฅ , ๐‘ฆ) = (๐‘ฆ , ๐‘ฅ + ๐‘ฆ) . (a) Show that ๐‘‡ ๐‘› (0 , 1) = (๐น ๐‘› , ๐น ๐‘› + 1 ) for each nonnegative integer ๐‘› . (b) Find the eigenvalues of ๐‘‡ . (c) Find a basis of ๐‘ 2 consisting of eigenvectors of ๐‘‡ . (d) Use the solution to (c) to compute ๐‘‡ ๐‘› (0 , 1) . Conclude that ๐น ๐‘› = 1 โˆš 5[(1 + โˆš 5 2 ) ๐‘› โˆ’ (1 โˆ’ โˆš 5 2 ) ๐‘› ] for each nonnegative integer ๐‘› . (e) Use (d) to conclude that if ๐‘› is a nonnegative integer, then the Fibonacci number ๐น ๐‘› is the integer that is closest to 1 โˆš 5( 1 + โˆš 5 2 ) ๐‘› . Each ๐น ๐‘› is a nonnegative integer, even though the right side of the formula in ( d ) does not look like an integer. The number 1 + โˆš 5 2 is called the golden ratio . 22 Suppose ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐ด is an ๐‘› -by- ๐‘› matrix that is the matrix of ๐‘‡ with respect to some basis of ๐‘‰ . Prove that if |๐ด ๐‘— , ๐‘— | > ๐‘› โˆ‘ ๐‘˜=1๐‘˜โ‰ ๐‘— |๐ด ๐‘— , ๐‘˜ | for each ๐‘— โˆˆ {1 , โ€ฆ , ๐‘›} , then ๐‘‡ is invertible. This exercise states that if the diagonal entries of the matrix of ๐‘‡ are large compared to the nondiagonal entries, then ๐‘‡ is invertible. 23 Suppose the definition of the Gershgorin disks is changed so that the radius of the ๐‘˜ th disk is the sum of the absolute values of the entries in column (instead of row) ๐‘˜ of ๐ด , excluding the diagonal entry. Show that the Gershgorin disk theorem (5.67) still holds with this changed definition. Section 5E Commuting Operators 175 5E Commuting Operators 5.71 definition: commute โ€ข Two operators ๐‘† and ๐‘‡ on the same vector space commute if ๐‘†๐‘‡ = ๐‘‡๐‘† . โ€ข Two square matrices ๐ด and ๐ต of the same size commute if ๐ด๐ต = ๐ต๐ด . For example, if ๐‘‡ is an operator and ๐‘ , ๐‘ž โˆˆ ๐’ซ (๐…) , then ๐‘(๐‘‡) and ๐‘ž(๐‘‡) commute [ see 5.17(b) ] . As another example, if ๐ผ is the identity operator on ๐‘‰ , then ๐ผ commutes with every operator on ๐‘‰ . 5.72 example: partial differentiation operators commute Suppose ๐‘š is a nonnegative integer. Let ๐’ซ ๐‘š (๐‘ 2 ) denote the real vector space of polynomials (with real coefficients) in two real variables and of degree at most ๐‘š , with the usual operations of addition and scalar multiplication of real-valued functions. Thus the elements of ๐’ซ ๐‘š (๐‘ 2 ) are functions ๐‘ on ๐‘ 2 of the form 5.73 ๐‘ = โˆ‘ ๐‘— + ๐‘˜โ‰ค๐‘š ๐‘Ž ๐‘— , ๐‘˜ ๐‘ฅ ๐‘— ๐‘ฆ ๐‘˜ , where the indices ๐‘— and ๐‘˜ take on all nonnegative integer values such that ๐‘— + ๐‘˜ โ‰ค ๐‘š , each ๐‘Ž ๐‘— , ๐‘˜ is in ๐‘ , and ๐‘ฅ ๐‘— ๐‘ฆ ๐‘˜ denotes the function on ๐‘ 2 defined by (๐‘ฅ , ๐‘ฆ) โ†ฆ ๐‘ฅ ๐‘— ๐‘ฆ ๐‘˜ . Define operators ๐ท ๐‘ฅ , ๐ท ๐‘ฆ โˆˆ โ„’ ( ๐’ซ ๐‘š (๐‘ 2 )) by ๐ท ๐‘ฅ ๐‘ = ๐œ•๐‘ ๐œ•๐‘ฅ = โˆ‘ ๐‘— + ๐‘˜โ‰ค๐‘š ๐‘—๐‘Ž ๐‘— , ๐‘˜ ๐‘ฅ ๐‘—โˆ’1 ๐‘ฆ ๐‘˜ and ๐ท ๐‘ฆ ๐‘ = ๐œ•๐‘ ๐œ•๐‘ฆ = โˆ‘ ๐‘— + ๐‘˜โ‰ค๐‘š ๐‘˜๐‘Ž ๐‘— , ๐‘˜ ๐‘ฅ ๐‘— ๐‘ฆ ๐‘˜โˆ’1 , where ๐‘ is as in 5.73. The operators ๐ท ๐‘ฅ and ๐ท ๐‘ฆ are called partial differentiation operators because each of these operators differentiates with respect to one of the variables while pretending that the other variable is a constant. The operators ๐ท ๐‘ฅ and ๐ท ๐‘ฆ commute because if ๐‘ is as in 5.73, then (๐ท ๐‘ฅ ๐ท ๐‘ฆ )๐‘ = โˆ‘ ๐‘— + ๐‘˜โ‰ค๐‘š ๐‘—๐‘˜๐‘Ž ๐‘— , ๐‘˜ ๐‘ฅ ๐‘—โˆ’1 ๐‘ฆ ๐‘˜โˆ’1 = (๐ท ๐‘ฆ ๐ท ๐‘ฅ )๐‘. The equation ๐ท ๐‘ฅ ๐ท ๐‘ฆ = ๐ท ๐‘ฆ ๐ท ๐‘ฅ on ๐’ซ ๐‘š (๐‘ 2 ) illustrates a more general result that the order of partial differentiation does not matter for nice functions. All 214,358,881 ( which equals 11 8 ) pairs of the 2 -by- 2 matrices under con- sideration were checked by a computer to discover that only 674,609 of these pairs of matrices commute. Commuting matrices are unusual. For example, there are 214,358,881 pairs of 2 -by- 2 matrices all of whose entries are integers in the interval [โˆ’5 , 5] . Only about 0.3% of these pairs of matrices commute. 176 Chapter 5 Eigenvalues and Eigenvectors The next result shows that two operators commute if and only if their matrices (with respect to the same basis) commute. 5.74 commuting operators correspond to commuting matrices Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) and ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . Then ๐‘† and ๐‘‡ commute if and only if โ„ณ (๐‘† , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) and โ„ณ (๐‘‡ , (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› )) commute. Proof We have ๐‘† and ๐‘‡ commute โŸบ ๐‘†๐‘‡ = ๐‘‡๐‘† โŸบ โ„ณ (๐‘†๐‘‡) = โ„ณ (๐‘‡๐‘†) โŸบ โ„ณ (๐‘†) โ„ณ (๐‘‡) = โ„ณ (๐‘‡) โ„ณ (๐‘†) โŸบ โ„ณ (๐‘†) and โ„ณ (๐‘‡) commute , as desired. The next result shows that if two operators commute, then every eigenspace for one operator is invariant under the other operator. This result, which we will use several times, is one of the main reasons why a pair of commuting operators behaves better than a pair of operators that does not commute. 5.75 eigenspace is invariant under commuting operator Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) commute and ๐œ† โˆˆ ๐… . Then ๐ธ(๐œ† , ๐‘†) is invariant under ๐‘‡ . Proof Suppose ๐‘ฃ โˆˆ ๐ธ(๐œ† , ๐‘†) . Then ๐‘†(๐‘‡๐‘ฃ) = (๐‘†๐‘‡)๐‘ฃ = (๐‘‡๐‘†)๐‘ฃ = ๐‘‡(๐‘†๐‘ฃ) = ๐‘‡(๐œ†๐‘ฃ) = ๐œ†๐‘‡๐‘ฃ. The equation above shows that ๐‘‡๐‘ฃ โˆˆ ๐ธ(๐œ† , ๐‘†) . Thus ๐ธ(๐œ† , ๐‘†) is invariant under ๐‘‡ . Suppose we have two operators, each of which is diagonalizable. If we want to do computations involving both operators (for example, involving their sum), then we want the two operators to be diagonalizable by the same basis, which according to the next result is possible when the two operators commute. 5.76 simultaneous diagonalizablity โŸบ commutativity Two diagonalizable operators on the same vector space have diagonal matrices with respect to the same basis if and only if the two operators commute. Proof First suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) have diagonal matrices with respect to the same basis. The product of two diagonal matrices of the same size is the diagonal matrix obtained by multiplying the corresponding elements of the two diagonals. Thus any two diagonal matrices of the same size commute. Thus ๐‘† and ๐‘‡ commute, by 5.74. Section 5E Commuting Operators 177 To prove the implication in the other direction, now suppose that ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) are diagonalizable operators that commute. Let ๐œ† 1 , โ€ฆ , ๐œ† ๐‘š denote the distinct eigenvalues of ๐‘† . Because ๐‘† is diagonalizable, 5.55(c) shows that 5.77 ๐‘‰ = ๐ธ(๐œ† 1 , ๐‘†) โŠ• โ‹ฏ โŠ• ๐ธ(๐œ† ๐‘š , ๐‘†). For each ๐‘˜ = 1 , โ€ฆ , ๐‘š , the subspace ๐ธ(๐œ† ๐‘˜ , ๐‘†) is invariant under ๐‘‡ (by 5.75). Because ๐‘‡ is diagonalizable, 5.65 implies that ๐‘‡| ๐ธ(๐œ† ๐‘˜ , ๐‘†) is diagonalizable for each ๐‘˜ . Hence for each ๐‘˜ = 1 , โ€ฆ , ๐‘š , there is a basis of ๐ธ(๐œ† ๐‘˜ , ๐‘†) consisting of eigenvectors of ๐‘‡ . Putting these bases together gives a basis of ๐‘‰ (because of 5.77), with each vector in this basis being an eigenvector of both ๐‘† and ๐‘‡ . Thus ๐‘† and ๐‘‡ both have diagonal matrices with respect to this basis, as desired. See Exercise 2 for an extension of the result above to more than two operators. Suppose ๐‘‰ is a finite-dimensional nonzero complex vector space. Then every operator on ๐‘‰ has an eigenvector (see 5.19). The next result shows that if two operators on ๐‘‰ commute, then there is a vector in ๐‘‰ that is an eigenvector for both operators (but the two commuting operators might not have a common eigenvalue). For an extension of the next result to more than two operators, see Exercise 9(a). 5.78 common eigenvector for commuting operators Every pair of commuting operators on a finite-dimensional nonzero complex vector space has a common eigenvector. Proof Suppose ๐‘‰ is a finite-dimensional nonzero complex vector space and ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) commute. Let ๐œ† be an eigenvalue of ๐‘† (5.19 tells us that ๐‘† does indeed have an eigenvalue). Thus ๐ธ(๐œ† , ๐‘†) โ‰  {0} . Also, ๐ธ(๐œ† , ๐‘†) is invariant under ๐‘‡ (by 5.75). Thus ๐‘‡| ๐ธ(๐œ† , ๐‘†) has an eigenvector (again using 5.19), which is an eigenvector for both ๐‘† and ๐‘‡ , completing the proof. 5.79 example: common eigenvector for partial differentiation operators Let ๐’ซ ๐‘š (๐‘ 2 ) be as in Example 5.72 and let ๐ท ๐‘ฅ , ๐ท ๐‘ฆ โˆˆ โ„’ ( ๐’ซ ๐‘š (๐‘ 2 )) be the commuting partial differentiation operators in that example. As you can verify, 0 is the only eigenvalue of each of these operators. Also ๐ธ(0 , ๐ท ๐‘ฅ ) = { ๐‘š โˆ‘ ๐‘˜=0 ๐‘Ž ๐‘˜ ๐‘ฆ ๐‘˜ โˆถ ๐‘Ž 0 , โ€ฆ , ๐‘Ž ๐‘š โˆˆ ๐‘} , ๐ธ(0 , ๐ท ๐‘ฆ ) = { ๐‘š โˆ‘ ๐‘—=0 ๐‘ ๐‘— ๐‘ฅ ๐‘— โˆถ ๐‘ 0 , โ€ฆ , ๐‘ ๐‘š โˆˆ ๐‘}. The intersection of these two eigenspaces is the set of common eigenvectors of the two operators. Because ๐ธ(0 , ๐ท ๐‘ฅ ) โˆฉ ๐ธ(0 , ๐ท ๐‘ฆ ) is the set of constant functions, we see that ๐ท ๐‘ฅ and ๐ท ๐‘ฆ indeed have a common eigenvector, as promised by 5.78. 178 Chapter 5 Eigenvalues and Eigenvectors The next result extends 5.47 (the existence of a basis that gives an upper- triangular matrix) to two commuting operators. 5.80 commuting operators are simultaneously upper triangularizable Suppose ๐‘‰ is a finite-dimensional complex vector space and ๐‘† , ๐‘‡ are commuting operators on ๐‘‰ . Then there is a basis of ๐‘‰ with respect to which both ๐‘† and ๐‘‡ have upper-triangular matrices. Proof Let ๐‘› = dim ๐‘‰ . We will use induction on ๐‘› . The desired result holds if ๐‘› = 1 because all 1 -by- 1 matrices are upper triangular. Now suppose ๐‘› > 1 and the desired result holds for all complex vector spaces whose dimension is ๐‘› โˆ’ 1 . Let ๐‘ฃ 1 be any common eigenvector of ๐‘† and ๐‘‡ (using 5.78). Hence ๐‘†๐‘ฃ 1 โˆˆ span (๐‘ฃ 1 ) and ๐‘‡๐‘ฃ 1 โˆˆ span (๐‘ฃ 1 ) . Let ๐‘Š be a subspace of ๐‘‰ such that ๐‘‰ = span (๐‘ฃ 1 ) โŠ• ๐‘Š ; see 2.33 for the existence of ๐‘Š . Define a linear map ๐‘ƒ โˆถ ๐‘‰ โ†’ ๐‘Š by ๐‘ƒ(๐‘Ž๐‘ฃ 1 + ๐‘ค) = ๐‘ค for each ๐‘Ž โˆˆ ๐‚ and each ๐‘ค โˆˆ ๐‘Š . Define ฬ‚๐‘† , ฬ‚๐‘‡ โˆˆ โ„’ (๐‘Š) by ฬ‚๐‘†๐‘ค = ๐‘ƒ(๐‘†๐‘ค) and ฬ‚๐‘‡๐‘ค = ๐‘ƒ(๐‘‡๐‘ค) for each ๐‘ค โˆˆ ๐‘Š . To apply our induction hypothesis to ฬ‚๐‘† and ฬ‚๐‘‡ , we must first show that these two operators on ๐‘Š commute. To do this, suppose ๐‘ค โˆˆ ๐‘Š . Then there exists ๐‘Ž โˆˆ ๐‚ such that ( ฬ‚๐‘† ฬ‚๐‘‡)๐‘ค = ฬ‚๐‘†(๐‘ƒ(๐‘‡๐‘ค)) = ฬ‚๐‘†(๐‘‡๐‘ค โˆ’ ๐‘Ž๐‘ฃ 1 ) = ๐‘ƒ(๐‘†(๐‘‡๐‘ค โˆ’ ๐‘Ž๐‘ฃ 1 )) = ๐‘ƒ((๐‘†๐‘‡)๐‘ค) , where the last equality holds because ๐‘ฃ 1 is an eigenvector of ๐‘† and ๐‘ƒ๐‘ฃ 1 = 0 . Similarly, ( ฬ‚๐‘‡ ฬ‚๐‘†)๐‘ค = ๐‘ƒ((๐‘‡๐‘†)๐‘ค). Because the operators ๐‘† and ๐‘‡ commute, the last two displayed equations show that ( ฬ‚๐‘† ฬ‚๐‘‡)๐‘ค = ( ฬ‚๐‘‡ ฬ‚๐‘†)๐‘ค . Hence ฬ‚๐‘† and ฬ‚๐‘‡ commute. Thus we can use our induction hypothesis to state that there exists a basis ๐‘ฃ 2 , โ€ฆ , ๐‘ฃ ๐‘› of ๐‘Š such that ฬ‚๐‘† and ฬ‚๐‘‡ both have upper-triangular matrices with respect to this basis. The list ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› is a basis of ๐‘‰ . If ๐‘˜ โˆˆ {2 , โ€ฆ , ๐‘›} , then there exist ๐‘Ž ๐‘˜ , ๐‘ ๐‘˜ โˆˆ ๐‚ such that ๐‘†๐‘ฃ ๐‘˜ = ๐‘Ž ๐‘˜ ๐‘ฃ 1 + ฬ‚๐‘†๐‘ฃ ๐‘˜ and ๐‘‡๐‘ฃ ๐‘˜ = ๐‘ ๐‘˜ ๐‘ฃ 1 + ฬ‚๐‘‡๐‘ฃ ๐‘˜ . Because ฬ‚๐‘† and ฬ‚๐‘‡ have upper-triangular matrices with respect to ๐‘ฃ 2 , โ€ฆ , ๐‘ฃ ๐‘› , we know that ฬ‚๐‘†๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 2 , โ€ฆ , ๐‘ฃ ๐‘˜ ) and ฬ‚๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 2 , โ€ฆ , ๐‘ฃ ๐‘˜ ) . Hence the equations above imply that ๐‘†๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ) and ๐‘‡๐‘ฃ ๐‘˜ โˆˆ span (๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘˜ ). Thus ๐‘† and ๐‘‡ have upper-triangular matrices with respect to ๐‘ฃ 1 , โ€ฆ , ๐‘ฃ ๐‘› , as desired. Exercise 9(b) extends the result above to more than two operators. Section 5E Commuting Operators 179 In general, it is not possible to determine the eigenvalues of the sum or product of two operators from the eigenvalues of the two operators. However, the next result shows that something nice happens when the two operators commute. 5.81 eigenvalues of sum and product of commuting operators Suppose ๐‘‰ is a finite-dimensional complex vector space and ๐‘† , ๐‘‡ are commut- ing operators on ๐‘‰ . Then โ€ข every eigenvalue of ๐‘† + ๐‘‡ is an eigenvalue of ๐‘† plus an eigenvalue of ๐‘‡ , โ€ข every eigenvalue of ๐‘†๐‘‡ is an eigenvalue of ๐‘† times an eigenvalue of ๐‘‡ . Proof There is a basis of ๐‘‰ with respect to which both ๐‘† and ๐‘‡ have upper- triangular matrices (by 5.80). With respect to that basis, โ„ณ (๐‘† + ๐‘‡) = โ„ณ (๐‘†) + โ„ณ (๐‘‡) and โ„ณ (๐‘†๐‘‡) = โ„ณ (๐‘†) โ„ณ (๐‘‡) , as stated in 3.35 and 3.43. The definition of matrix addition shows that each entry on the diagonal of โ„ณ (๐‘† + ๐‘‡) equals the sum of the corresponding entries on the diagonals of โ„ณ (๐‘†) and โ„ณ (๐‘‡) . Similarly, because โ„ณ (๐‘†) and โ„ณ (๐‘‡) are upper-triangular matrices, the definition of matrix multiplication shows that each entry on the diagonal of โ„ณ (๐‘†๐‘‡) equals the product of the corresponding entries on the diagonals of โ„ณ (๐‘†) and โ„ณ (๐‘‡) . Furthermore, โ„ณ (๐‘† + ๐‘‡) and โ„ณ (๐‘†๐‘‡) are upper-triangular matrices (see Exercise 2 in Section 5C). Every entry on the diagonal of โ„ณ (๐‘†) is an eigenvalue of ๐‘† , and every entry on the diagonal of โ„ณ (๐‘‡) is an eigenvalue of ๐‘‡ (by 5.41). Every eigenvalue of ๐‘† + ๐‘‡ is on the diagonal of โ„ณ (๐‘† + ๐‘‡) , and every eigenvalue of ๐‘†๐‘‡ is on the diagonal of โ„ณ (๐‘†๐‘‡) (these assertions follow from 5.41). Putting all this together, we conclude that every eigenvalue of ๐‘† + ๐‘‡ is an eigenvalue of ๐‘† plus an eigenvalue of ๐‘‡ , and every eigenvalue of ๐‘†๐‘‡ is an eigenvalue of ๐‘† times an eigenvalue of ๐‘‡ . Exercises 5E 1 Give an example of two commuting operators ๐‘† , ๐‘‡ on ๐… 4 such that there is a subspace of ๐… 4 that is invariant under ๐‘† but not under ๐‘‡ and there is a subspace of ๐… 4 that is invariant under ๐‘‡ but not under ๐‘† . 2 Suppose โ„ฐ is a subset of โ„’ (๐‘‰) and every element of โ„ฐ is diagonalizable. Prove that there exists a basis of ๐‘‰ with respect to which every element of โ„ฐ has a diagonal matrix if and only if every pair of elements of โ„ฐ commutes. This exercise extends 5.76, which considers the case in which โ„ฐ contains only two elements. For this exercise, โ„ฐ may contain any number of elements, and โ„ฐ may even be an infinite set. 180 Chapter 5 Eigenvalues and Eigenvectors 3 Suppose ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) are such that ๐‘†๐‘‡ = ๐‘‡๐‘† . Suppose ๐‘ โˆˆ ๐’ซ (๐…) . (a) Prove that null ๐‘(๐‘†) is invariant under ๐‘‡ . (b) Prove that range ๐‘(๐‘†) is invariant under ๐‘‡ . See 5.18 for the special case ๐‘† = ๐‘‡ . 4 Prove or give a counterexample: If ๐ด is a diagonal matrix and ๐ต is an upper-triangular matrix of the same size as ๐ด , then ๐ด and ๐ต commute. 5 Prove that a pair of operators on a finite-dimensional vector space commute if and only if their dual operators commute. See 3.118 for the definition of the dual of an operator. 6 Suppose ๐‘‰ is a finite-dimensional complex vector space and ๐‘† , ๐‘‡ โˆˆ โ„’ (๐‘‰) commute. Prove that there exist ๐›ผ , ๐œ† โˆˆ ๐‚ such that range (๐‘† โˆ’ ๐›ผ๐ผ) + range (๐‘‡ โˆ’ ๐œ†๐ผ) โ‰  ๐‘‰. 7 Suppose ๐‘‰ is a complex vector space, ๐‘† โˆˆ โ„’ (๐‘‰) is diagonalizable, and ๐‘‡ โˆˆ โ„’ (๐‘‰) commutes with ๐‘† . Prove that there is a basis of ๐‘‰ such that ๐‘† has a diagonal matrix with respect to this basis and ๐‘‡ has an upper-triangular matrix with respect to this basis. 8 Suppose ๐‘š = 3 in Example 5.72 and ๐ท ๐‘ฅ , ๐ท ๐‘ฆ are the commuting partial differentiation operators on ๐’ซ 3 (๐‘ 2 ) from that example. Find a basis of ๐’ซ 3 (๐‘ 2 ) with respect to which ๐ท ๐‘ฅ and ๐ท ๐‘ฆ each have an upper-triangular matrix. 9 Suppose ๐‘‰ is a finite-dimensional nonzero complex vector space. Suppose that โ„ฐ โІ โ„’ (๐‘‰) is such that ๐‘† and ๐‘‡ commute for all ๐‘† , ๐‘‡ โˆˆ โ„ฐ . (a) Prove that there is a vector in ๐‘‰ that is an eigenvector for every element of โ„ฐ . (b) Prove that there is a basis of ๐‘‰ with respect to which every element of โ„ฐ has an upper-triangular matrix. This exercise extends 5.78 and 5.80, which consider the case in which โ„ฐ contains only two elements. For this exercise, โ„ฐ may contain any number of elements, and โ„ฐ may even be an infinite set. 10 Give an example of two commuting operators ๐‘† , ๐‘‡ on a finite-dimensional real vector space such that ๐‘† + ๐‘‡ has a eigenvalue that does not equal an eigenvalue of ๐‘† plus an eigenvalue of ๐‘‡ and ๐‘†๐‘‡ has a eigenvalue that does not equal an eigenvalue of ๐‘† times an eigenvalue of ๐‘‡ . This exercise shows that 5.81 does not hold on real vector spaces.