Chapter 6 Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition and scalar multiplication) of 𝐑 2 and 𝐑 3 . We ignored geometric features such as the notions of length and angle. These ideas are embedded in the concept of inner products, which we will investigate in this chapter. Every inner product induces a norm, which you can think of as a length. This norm satisfies key properties such as the Pythagorean theorem, the triangle inequality, the parallelogram equality, and the Cauchy–Schwarz inequality. The notion of perpendicular vectors in Euclidean geometry gets renamed to orthogonal vectors in the context of an inner product space. We will see that orthonormal bases are tremendously useful in inner product spaces. The Gram– Schmidt procedure constructs such bases. This chapter will conclude by putting together these tools to solve minimization problems. standing assumptions for this chapter β€’ 𝐅 denotes 𝐑 or 𝐂 . β€’ 𝑉 and π‘Š denote vector spaces over 𝐅 . M a tt he w P e t r o ff CC BY - SA The George Peabody Library, now part of Johns Hopkins University, opened while James Sylvester ( 1814–1897 ) was the university’s first mathematics professor. Sylvester’s publications include the first use of the word matrix in mathematics. 181 182 Chapter 6 Inner Product Spaces 6A Inner Products and Norms Inner Products This vector 𝑣 has norm √ π‘Ž 2 + 𝑏 2 . To motivate the concept of inner product, think of vectors in 𝐑 2 and 𝐑 3 as arrows with initial point at the origin. The length of a vector 𝑣 in 𝐑 2 or 𝐑 3 is called the norm of 𝑣 and is denoted by ‖𝑣‖ . Thus for 𝑣 = (π‘Ž , 𝑏) ∈ 𝐑 2 , we have ‖𝑣‖ = √ π‘Ž 2 + 𝑏 2 . Similarly, if 𝑣 = (π‘Ž , 𝑏 , 𝑐) ∈ 𝐑 3 , then ‖𝑣‖ = √ π‘Ž 2 + 𝑏 2 + 𝑐 2 . Even though we cannot draw pictures in higher dimensions, the generalization to 𝐑 𝑛 is easy: we define the norm of π‘₯ = (π‘₯ 1 , … , π‘₯ 𝑛 ) ∈ 𝐑 𝑛 by β€–π‘₯β€– = √ π‘₯ 12 + β‹― + π‘₯ 𝑛2 . The norm is not linear on 𝐑 𝑛 . To inject linearity into the discussion, we introduce the dot product. 6.1 definition: dot product For π‘₯ , 𝑦 ∈ 𝐑 𝑛 , the dot product of π‘₯ and 𝑦 , denoted by π‘₯ β‹… 𝑦 , is defined by π‘₯ β‹… 𝑦 = π‘₯ 1 𝑦 1 + β‹― + π‘₯ 𝑛 𝑦 𝑛 , where π‘₯ = (π‘₯ 1 , … , π‘₯ 𝑛 ) and 𝑦 = (𝑦 1 , … , 𝑦 𝑛 ) . If we think of a vector as a point instead of as an arrow, then β€–π‘₯β€– should be interpreted to mean the distance from the origin to the point π‘₯ . The dot product of two vectors in 𝐑 𝑛 is a number, not a vector. Notice that π‘₯ β‹… π‘₯ = β€–π‘₯β€– 2 for all π‘₯ ∈ 𝐑 𝑛 . Furthermore, the dot product on 𝐑 𝑛 has the following properties. β€’ π‘₯ β‹… π‘₯ β‰₯ 0 for all π‘₯ ∈ 𝐑 𝑛 . β€’ π‘₯ β‹… π‘₯ = 0 if and only if π‘₯ = 0 . β€’ For 𝑦 ∈ 𝐑 𝑛 fixed, the map from 𝐑 𝑛 to 𝐑 that sends π‘₯ ∈ 𝐑 𝑛 to π‘₯ β‹… 𝑦 is linear. β€’ π‘₯ β‹… 𝑦 = 𝑦 β‹… π‘₯ for all π‘₯ , 𝑦 ∈ 𝐑 𝑛 . An inner product is a generalization of the dot product. At this point you may be tempted to guess that an inner product is defined by abstracting the properties of the dot product discussed in the last paragraph. For real vector spaces, that guess is correct. However, so that we can make a definition that will be useful for both real and complex vector spaces, we need to examine the complex case before making the definition. Section 6A Inner Products and Norms 183 Recall that if πœ† = π‘Ž + 𝑏𝑖 , where π‘Ž , 𝑏 ∈ 𝐑 , then β€’ the absolute value of πœ† , denoted by |πœ†| , is defined by |πœ†| = √ π‘Ž 2 + 𝑏 2 ; β€’ the complex conjugate of πœ† , denoted by πœ† , is defined by πœ† = π‘Ž βˆ’ 𝑏𝑖 ; β€’ |πœ†| 2 = πœ†πœ† . See Chapter 4 for the definitions and the basic properties of the absolute value and complex conjugate. For 𝑧 = (𝑧 1 , … , 𝑧 𝑛 ) ∈ 𝐂 𝑛 , we define the norm of 𝑧 by ‖𝑧‖ = √ |𝑧 1 | 2 + β‹― + |𝑧 𝑛 | 2 . The absolute values are needed because we want ‖𝑧‖ to be a nonnegative number. Note that ‖𝑧‖ 2 = 𝑧 1 𝑧 1 + β‹― + 𝑧 𝑛 𝑧 𝑛 . We want to think of ‖𝑧‖ 2 as the inner product of 𝑧 with itself, as we did in 𝐑 𝑛 . The equation above thus suggests that the inner product of the vector 𝑀 = (𝑀 1 , … , 𝑀 𝑛 ) ∈ 𝐂 𝑛 with 𝑧 should equal 𝑀 1 𝑧 1 + β‹― + 𝑀 𝑛 𝑧 𝑛 . If the roles of the 𝑀 and 𝑧 were interchanged, the expression above would be replaced with its complex conjugate. Thus we should expect that the inner product of 𝑀 with 𝑧 equals the complex conjugate of the inner product of 𝑧 with 𝑀 . With that motivation, we are now ready to define an inner product on 𝑉 , which may be a real or a complex vector space. One comment about the notation used in the next definition: β€’ For πœ† ∈ 𝐂 , the notation πœ† β‰₯ 0 means πœ† is real and nonnegative. 6.2 definition: inner product An inner product on 𝑉 is a function that takes each ordered pair (𝑒 , 𝑣) of elements of 𝑉 to a number βŸ¨π‘’ , π‘£βŸ© ∈ 𝐅 and has the following properties. positivity βŸ¨π‘£ , π‘£βŸ© β‰₯ 0 for all 𝑣 ∈ 𝑉 . definiteness βŸ¨π‘£ , π‘£βŸ© = 0 if and only if 𝑣 = 0 . additivity in first slot βŸ¨π‘’ + 𝑣 , π‘€βŸ© = βŸ¨π‘’ , π‘€βŸ© + βŸ¨π‘£ , π‘€βŸ© for all 𝑒 , 𝑣 , 𝑀 ∈ 𝑉 . homogeneity in first slot βŸ¨πœ†π‘’ , π‘£βŸ© = πœ†βŸ¨π‘’ , π‘£βŸ© for all πœ† ∈ 𝐅 and all 𝑒 , 𝑣 ∈ 𝑉 . conjugate symmetry βŸ¨π‘’ , π‘£βŸ© = βŸ¨π‘£ , π‘’βŸ© for all 𝑒 , 𝑣 ∈ 𝑉 . 184 Chapter 6 Inner Product Spaces Most mathematicians define inner products as above, but many physicists use a definition that requires homo- geneity in the second slot instead of the first slot. Every real number equals its complex conjugate. Thus if we are dealing with a real vector space, then in the last con- dition above we can dispense with the complex conjugate and simply state that βŸ¨π‘’ , π‘£βŸ© = βŸ¨π‘£ , π‘’βŸ© for all 𝑒 , 𝑣 ∈ 𝑉 . 6.3 example: inner products (a) The Euclidean inner product on 𝐅 𝑛 is defined by ⟨(𝑀 1 , … , 𝑀 𝑛 ) , (𝑧 1 , … , 𝑧 𝑛 )⟩ = 𝑀 1 𝑧 1 + β‹― + 𝑀 𝑛 𝑧 𝑛 for all (𝑀 1 , … , 𝑀 𝑛 ) , (𝑧 1 , … , 𝑧 𝑛 ) ∈ 𝐅 𝑛 . (b) If 𝑐 1 , … , 𝑐 𝑛 are positive numbers, then an inner product can be defined on 𝐅 𝑛 by ⟨(𝑀 1 , … , 𝑀 𝑛 ) , (𝑧 1 , … , 𝑧 𝑛 )⟩ = 𝑐 1 𝑀 1 𝑧 1 + β‹― + 𝑐 𝑛 𝑀 𝑛 𝑧 𝑛 for all (𝑀 1 , … , 𝑀 𝑛 ) , (𝑧 1 , … , 𝑧 𝑛 ) ∈ 𝐅 𝑛 . (c) An inner product can be defined on the vector space of continuous real-valued functions on the interval [βˆ’1 , 1] by ⟨ 𝑓 , π‘”βŸ© = ∫ 1 βˆ’1 𝑓 𝑔 for all 𝑓 , 𝑔 continuous real-valued functions on [βˆ’1 , 1] . (d) An inner product can be defined on 𝒫 (𝐑) by βŸ¨π‘ , π‘žβŸ© = 𝑝(0)π‘ž(0) + ∫ 1 βˆ’1 𝑝 β€² π‘ž β€² for all 𝑝 , π‘ž ∈ 𝒫 (𝐑) . (e) An inner product can be defined on 𝒫 (𝐑) by βŸ¨π‘ , π‘žβŸ© = ∫ ∞ 0 𝑝(π‘₯)π‘ž(π‘₯)𝑒 βˆ’π‘₯ 𝑑π‘₯ for all 𝑝 , π‘ž ∈ 𝒫 (𝐑) . 6.4 definition: inner product space An inner product space is a vector space 𝑉 along with an inner product on 𝑉 . The most important example of an inner product space is 𝐅 𝑛 with the Euclidean inner product given by (a) in the example above. When 𝐅 𝑛 is referred to as an inner product space, you should assume that the inner product is the Euclidean inner product unless explicitly told otherwise. Section 6A Inner Products and Norms 185 So that we do not have to keep repeating the hypothesis that 𝑉 and π‘Š are inner product spaces, we make the following assumption. 6.5 notation: 𝑉 , π‘Š For the rest of this chapter and the next chapter, 𝑉 and π‘Š denote inner product spaces over 𝐅 . Note the slight abuse of language here. An inner product space is a vector space along with an inner product on that vector space. When we say that a vector space 𝑉 is an inner product space, we are also thinking that an inner product on 𝑉 is lurking nearby or is clear from the context ( or is the Euclidean inner product if the vector space is 𝐅 𝑛 ) . 6.6 basic properties of an inner product (a) For each fixed 𝑣 ∈ 𝑉 , the function that takes 𝑒 ∈ 𝑉 to βŸ¨π‘’ , π‘£βŸ© is a linear map from 𝑉 to 𝐅 . (b) ⟨0 , π‘£βŸ© = 0 for every 𝑣 ∈ 𝑉 . (c) βŸ¨π‘£ , 0⟩ = 0 for every 𝑣 ∈ 𝑉 . (d) βŸ¨π‘’ , 𝑣 + π‘€βŸ© = βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘’ , π‘€βŸ© for all 𝑒 , 𝑣 , 𝑀 ∈ 𝑉 . (e) βŸ¨π‘’ , πœ†π‘£βŸ© = πœ†βŸ¨π‘’ , π‘£βŸ© for all πœ† ∈ 𝐅 and all 𝑒 , 𝑣 ∈ 𝑉 . Proof (a) For 𝑣 ∈ 𝑉 , the linearity of 𝑒 ↦ βŸ¨π‘’ , π‘£βŸ© follows from the conditions of additivity and homogeneity in the first slot in the definition of an inner product. (b) Every linear map takes 0 to 0 . Thus (b) follows from (a). (c) If 𝑣 ∈ 𝑉 , then the conjugate symmetry property in the definition of an inner product and (b) show that βŸ¨π‘£ , 0⟩ = ⟨0 , π‘£βŸ© = 0 = 0 . (d) Suppose 𝑒 , 𝑣 , 𝑀 ∈ 𝑉 . Then βŸ¨π‘’ , 𝑣 + π‘€βŸ© = βŸ¨π‘£ + 𝑀 , π‘’βŸ© = βŸ¨π‘£ , π‘’βŸ© + βŸ¨π‘€ , π‘’βŸ© = βŸ¨π‘£ , π‘’βŸ© + βŸ¨π‘€ , π‘’βŸ© = βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘’ , π‘€βŸ©. (e) Suppose πœ† ∈ 𝐅 and 𝑒 , 𝑣 ∈ 𝑉 . Then βŸ¨π‘’ , πœ†π‘£βŸ© = βŸ¨πœ†π‘£ , π‘’βŸ© = πœ†βŸ¨π‘£ , π‘’βŸ© = πœ† βŸ¨π‘£ , π‘’βŸ© = πœ†βŸ¨π‘’ , π‘£βŸ©. 186 Chapter 6 Inner Product Spaces Norms Our motivation for defining inner products came initially from the norms of vectors on 𝐑 2 and 𝐑 3 . Now we see that each inner product determines a norm. 6.7 definition: norm, ‖𝑣‖ For 𝑣 ∈ 𝑉 , the norm of 𝑣 , denoted by ‖𝑣‖ , is defined by ‖𝑣‖ = √ βŸ¨π‘£ , π‘£βŸ©. 6.8 example: norms (a) If (𝑧 1 , … , 𝑧 𝑛 ) ∈ 𝐅 𝑛 (with the Euclidean inner product), then β€–(𝑧 1 , … , 𝑧 𝑛 )β€– = √ |𝑧 1 | 2 + β‹― + |𝑧 𝑛 | 2 . (b) For 𝑓 in the vector space of continuous real-valued functions on [βˆ’1 , 1] and with inner product given as in 6.3(c), we have β€– 𝑓 β€– = √ ∫ 1 βˆ’1 𝑓 2 . 6.9 basic properties of the norm Suppose 𝑣 ∈ 𝑉 . (a) ‖𝑣‖ = 0 if and only if 𝑣 = 0 . (b) β€–πœ†π‘£β€– = |πœ†| ‖𝑣‖ for all πœ† ∈ 𝐅 . Proof (a) The desired result holds because βŸ¨π‘£ , π‘£βŸ© = 0 if and only if 𝑣 = 0 . (b) Suppose πœ† ∈ 𝐅 . Then β€–πœ†π‘£β€– 2 = βŸ¨πœ†π‘£ , πœ†π‘£βŸ© = πœ†βŸ¨π‘£ , πœ†π‘£βŸ© = πœ†πœ†βŸ¨π‘£ , π‘£βŸ© = |πœ†| 2 ‖𝑣‖ 2 . Taking square roots now gives the desired equality. The proof of (b) in the result above illustrates a general principle: working with norms squared is usually easier than working directly with norms. Section 6A Inner Products and Norms 187 Now we come to a crucial definition. 6.10 definition: orthogonal Two vectors 𝑒 , 𝑣 ∈ 𝑉 are called orthogonal if βŸ¨π‘’ , π‘£βŸ© = 0 . The word orthogonal comes from the Greek word orthogonios , which means right-angled. In the definition above, the order of the two vectors does not matter, because βŸ¨π‘’ , π‘£βŸ© = 0 if and only if βŸ¨π‘£ , π‘’βŸ© = 0 . In- stead of saying 𝑒 and 𝑣 are orthogonal, sometimes we say 𝑒 is orthogonal to 𝑣 . Exercise 15 asks you to prove that if 𝑒 , 𝑣 are nonzero vectors in 𝐑 2 , then βŸ¨π‘’ , π‘£βŸ© = ‖𝑒‖ ‖𝑣‖ cos πœƒ , where πœƒ is the angle between 𝑒 and 𝑣 (thinking of 𝑒 and 𝑣 as arrows with initial point at the origin). Thus two nonzero vectors in 𝐑 2 are orthogonal (with respect to the Euclidean inner product) if and only if the cosine of the angle between them is 0 , which happens if and only if the vectors are perpendicular in the usual sense of plane geometry. Thus you can think of the word orthogonal as a fancy word meaning perpendicular . We begin our study of orthogonality with an easy result. 6.11 orthogonality and 0 (a) 0 is orthogonal to every vector in 𝑉 . (b) 0 is the only vector in 𝑉 that is orthogonal to itself. Proof (a) Recall that 6.6(b) states that ⟨0 , π‘£βŸ© = 0 for every 𝑣 ∈ 𝑉 . (b) If 𝑣 ∈ 𝑉 and βŸ¨π‘£ , π‘£βŸ© = 0 , then 𝑣 = 0 (by definition of inner product). For the special case 𝑉 = 𝐑 2 , the next theorem was known over 3,500 years ago in Babylonia and then rediscovered and proved over 2,500 years ago in Greece. Of course, the proof below is not the original proof. 6.12 Pythagorean theorem Suppose 𝑒 , 𝑣 ∈ 𝑉 . If 𝑒 and 𝑣 are orthogonal, then ‖𝑒 + 𝑣‖ 2 = ‖𝑒‖ 2 + ‖𝑣‖ 2 . Proof Suppose βŸ¨π‘’ , π‘£βŸ© = 0 . Then ‖𝑒 + 𝑣‖ 2 = βŸ¨π‘’ + 𝑣 , 𝑒 + π‘£βŸ© = βŸ¨π‘’ , π‘’βŸ© + βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘£ , π‘’βŸ© + βŸ¨π‘£ , π‘£βŸ© = ‖𝑒‖ 2 + ‖𝑣‖ 2 . 188 Chapter 6 Inner Product Spaces Suppose 𝑒 , 𝑣 ∈ 𝑉 , with 𝑣 β‰  0 . We would like to write 𝑒 as a scalar multiple of 𝑣 plus a vector 𝑀 orthogonal to 𝑣 , as suggested in the picture here. An orthogonal decomposition: 𝑒 expressed as a scalar multiple of 𝑣 plus a vector orthogonal to 𝑣 . To discover how to write 𝑒 as a scalar multiple of 𝑣 plus a vector orthogonal to 𝑣 , let 𝑐 ∈ 𝐅 denote a scalar. Then 𝑒 = 𝑐𝑣 + (𝑒 βˆ’ 𝑐𝑣). Thus we need to choose 𝑐 so that 𝑣 is orthogonal to (𝑒 βˆ’ 𝑐𝑣) . Hence we want 0 = βŸ¨π‘’ βˆ’ 𝑐𝑣 , π‘£βŸ© = βŸ¨π‘’ , π‘£βŸ© βˆ’ 𝑐‖𝑣‖ 2 . The equation above shows that we should choose 𝑐 to be βŸ¨π‘’ , π‘£βŸ©/‖𝑣‖ 2 . Making this choice of 𝑐 , we can write 𝑒 = βŸ¨π‘’ , π‘£βŸ© ‖𝑣‖ 2 𝑣 + (𝑒 βˆ’ βŸ¨π‘’ , π‘£βŸ© ‖𝑣‖ 2 𝑣). As you should verify, the equation displayed above explicitly writes 𝑒 as a scalar multiple of 𝑣 plus a vector orthogonal to 𝑣 . Thus we have proved the following key result. 6.13 an orthogonal decomposition Suppose 𝑒 , 𝑣 ∈ 𝑉 , with 𝑣 β‰  0 . Set 𝑐 = βŸ¨π‘’ , π‘£βŸ© ‖𝑣‖ 2 and 𝑀 = 𝑒 βˆ’ βŸ¨π‘’ , π‘£βŸ© ‖𝑣‖ 2 𝑣 . Then 𝑒 = 𝑐𝑣 + 𝑀 and βŸ¨π‘€ , π‘£βŸ© = 0. The orthogonal decomposition 6.13 will be used in the proof of the Cauchy– Schwarz inequality, which is our next result and is one of the most important inequalities in mathematics. Section 6A Inner Products and Norms 189 6.14 Cauchy–Schwarz inequality Suppose 𝑒 , 𝑣 ∈ 𝑉 . Then |βŸ¨π‘’ , π‘£βŸ©| ≀ ‖𝑒‖ ‖𝑣‖. This inequality is an equality if and only if one of 𝑒 , 𝑣 is a scalar multiple of the other. Proof If 𝑣 = 0 , then both sides of the desired inequality equal 0 . Thus we can assume that 𝑣 β‰  0 . Consider the orthogonal decomposition 𝑒 = βŸ¨π‘’ , π‘£βŸ© ‖𝑣‖ 2 𝑣 + 𝑀 given by 6.13, where 𝑀 is orthogonal to 𝑣 . By the Pythagorean theorem, ‖𝑒‖ 2 = βˆ₯βŸ¨π‘’ , π‘£βŸ© ‖𝑣‖ 2 𝑣βˆ₯ 2 + ‖𝑀‖ 2 = βˆ£βŸ¨π‘’ , π‘£βŸ©βˆ£ 2 ‖𝑣‖ 2 + ‖𝑀‖ 2 β‰₯ βˆ£βŸ¨π‘’ , π‘£βŸ©βˆ£ 2 ‖𝑣‖ 2 . 6.15 Multiplying both sides of this inequality by ‖𝑣‖ 2 and then taking square roots gives the desired inequality. Augustin-Louis Cauchy ( 1789–1857 ) proved 6.16 ( a ) in 1821. In 1859, Cauchy’s student Viktor Bunyakovsky ( 1804–1889 ) proved integral inequal- ities like the one in 6.16 ( b ) . A few decades later, similar discoveries by Hermann Schwarz ( 1843–1921 ) at- tracted more attention and led to the name of this inequality. The proof in the paragraph above shows that the Cauchy–Schwarz inequal- ity is an equality if and only if 6.15 is an equality. This happens if and only if 𝑀 = 0 . But 𝑀 = 0 if and only if 𝑒 is a multiple of 𝑣 (see 6.13). Thus the Cauchy–Schwarz inequality is an equal- ity if and only if 𝑒 is a scalar multiple of 𝑣 or 𝑣 is a scalar multiple of 𝑒 (or both; the phrasing has been chosen to cover cases in which either 𝑒 or 𝑣 equals 0 ). 6.16 example: Cauchy–Schwarz inequality (a) If π‘₯ 1 , … , π‘₯ 𝑛 , 𝑦 1 , … , 𝑦 𝑛 ∈ 𝐑 , then (π‘₯ 1 𝑦 1 + β‹― + π‘₯ 𝑛 𝑦 𝑛 ) 2 ≀ (π‘₯ 12 + β‹― + π‘₯ 𝑛2 )(𝑦 12 + β‹― + 𝑦 𝑛2 ) , as follows from applying the Cauchy–Schwarz inequality to the vectors (π‘₯ 1 , … , π‘₯ 𝑛 ) , (𝑦 1 , … , 𝑦 𝑛 ) ∈ 𝐑 𝑛 , using the usual Euclidean inner product. 190 Chapter 6 Inner Product Spaces (b) If 𝑓 , 𝑔 are continuous real-valued functions on [βˆ’1 , 1] , then ∣ ∫ 1 βˆ’1 𝑓 𝑔 ∣ 2 ≀ ( ∫ 1 βˆ’1 𝑓 2 )( ∫ 1 βˆ’1 𝑔 2 ) , as follows from applying the Cauchy–Schwarz inequality to Example 6.3(c). In this triangle, the length of 𝑒 + 𝑣 is less than the length of 𝑒 plus the length of 𝑣 . The next result, called the triangle inequality, has the geometric interpretation that the length of each side of a triangle is less than the sum of the lengths of the other two sides. Note that the triangle inequality implies that the shortest polygonal path between two points is a single line segment (a polygonal path consists of line segments). 6.17 triangle inequality Suppose 𝑒 , 𝑣 ∈ 𝑉 . Then ‖𝑒 + 𝑣‖ ≀ ‖𝑒‖ + ‖𝑣‖. This inequality is an equality if and only if one of 𝑒 , 𝑣 is a nonnegative real multiple of the other. Proof We have ‖𝑒 + 𝑣‖ 2 = βŸ¨π‘’ + 𝑣 , 𝑒 + π‘£βŸ© = βŸ¨π‘’ , π‘’βŸ© + βŸ¨π‘£ , π‘£βŸ© + βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘£ , π‘’βŸ© = βŸ¨π‘’ , π‘’βŸ© + βŸ¨π‘£ , π‘£βŸ© + βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘’ , π‘£βŸ© = ‖𝑒‖ 2 + ‖𝑣‖ 2 + 2 Re βŸ¨π‘’ , π‘£βŸ© ≀ ‖𝑒‖ 2 + ‖𝑣‖ 2 + 2βˆ£βŸ¨π‘’ , π‘£βŸ©βˆ£ 6.18 ≀ ‖𝑒‖ 2 + ‖𝑣‖ 2 + 2‖𝑒‖ ‖𝑣‖ 6.19 = (‖𝑒‖ + ‖𝑣‖) 2 , where 6.19 follows from the Cauchy–Schwarz inequality (6.14). Taking square roots of both sides of the inequality above gives the desired inequality. The proof above shows that the triangle inequality is an equality if and only if we have equality in 6.18 and 6.19. Thus we have equality in the triangle inequality if and only if 6.20 βŸ¨π‘’ , π‘£βŸ© = ‖𝑒‖ ‖𝑣‖. If one of 𝑒 , 𝑣 is a nonnegative real multiple of the other, then 6.20 holds. Con- versely, suppose 6.20 holds. Then the condition for equality in the Cauchy– Schwarz inequality (6.14) implies that one of 𝑒 , 𝑣 is a scalar multiple of the other. This scalar must be a nonnegative real number, by 6.20, completing the proof. For the reverse triangle inequality, see Exercise 20. Section 6A Inner Products and Norms 191 The diagonals of this parallelogram are 𝑒 + 𝑣 and 𝑒 βˆ’ 𝑣 . The next result is called the parallel- ogram equality because of its geometric interpretation: in every parallelogram, the sum of the squares of the lengths of the diagonals equals the sum of the squares of the lengths of the four sides. Note that the proof here is more straightforward than the usual proof in Euclidean geometry. 6.21 parallelogram equality Suppose 𝑒 , 𝑣 ∈ 𝑉 . Then ‖𝑒 + 𝑣‖ 2 + ‖𝑒 βˆ’ 𝑣‖ 2 = 2(‖𝑒‖ 2 + ‖𝑣‖ 2 ). Proof We have ‖𝑒 + 𝑣‖ 2 + ‖𝑒 βˆ’ 𝑣‖ 2 = βŸ¨π‘’ + 𝑣 , 𝑒 + π‘£βŸ© + βŸ¨π‘’ βˆ’ 𝑣 , 𝑒 βˆ’ π‘£βŸ© = ‖𝑒‖ 2 + ‖𝑣‖ 2 + βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘£ , π‘’βŸ© + ‖𝑒‖ 2 + ‖𝑣‖ 2 βˆ’ βŸ¨π‘’ , π‘£βŸ© βˆ’ βŸ¨π‘£ , π‘’βŸ© = 2(‖𝑒‖ 2 + ‖𝑣‖ 2 ) , as desired. Exercises 6A 1 Prove or give a counterexample: If 𝑣 1 , … , 𝑣 π‘š ∈ 𝑉 , then π‘š βˆ‘ 𝑗 = 1 π‘š βˆ‘ π‘˜=1 βŸ¨π‘£ 𝑗 , 𝑣 π‘˜ ⟩ β‰₯ 0. 2 Suppose 𝑆 ∈ β„’ (𝑉) . Define βŸ¨β‹… , β‹…βŸ© 1 by βŸ¨π‘’ , π‘£βŸ© 1 = βŸ¨π‘†π‘’ , π‘†π‘£βŸ© for all 𝑒 , 𝑣 ∈ 𝑉 . Show that βŸ¨β‹… , β‹…βŸ© 1 is an inner product on 𝑉 if and only if 𝑆 is injective. 3 (a) Show that the function taking an ordered pair ((π‘₯ 1 , π‘₯ 2 ) , (𝑦 1 , 𝑦 2 )) of elements of 𝐑 2 to |π‘₯ 1 𝑦 1 | + |π‘₯ 2 𝑦 2 | is not an inner product on 𝐑 2 . (b) Show that the function taking an ordered pair ((π‘₯ 1 , π‘₯ 2 , π‘₯ 3 ) , (𝑦 1 , 𝑦 2 , 𝑦 3 )) of elements of 𝐑 3 to π‘₯ 1 𝑦 1 + π‘₯ 3 𝑦 3 is not an inner product on 𝐑 3 . 4 Suppose 𝑇 ∈ β„’ (𝑉) is such that ‖𝑇𝑣‖ ≀ ‖𝑣‖ for every 𝑣 ∈ 𝑉 . Prove that 𝑇 βˆ’ √ 2 𝐼 is injective. Annotated Entity: ID: 205 Spans: True Boxes: True Text: 192 Chapter 6 Inner Product Spaces 5 Suppose 𝑉 is a real inner product space. (a) Show that βŸ¨π‘’ + 𝑣 , 𝑒 βˆ’ π‘£βŸ© = ‖𝑒‖ 2 βˆ’ ‖𝑣‖ 2 for every 𝑒 , 𝑣 ∈ 𝑉 . (b) Show that if 𝑒 , 𝑣 ∈ 𝑉 have the same norm, then 𝑒 + 𝑣 is orthogonal to 𝑒 βˆ’ 𝑣 . (c) Use (b) to show that the diagonals of a rhombus are perpendicular to each other. 6 Suppose 𝑒 , 𝑣 ∈ 𝑉 . Prove that βŸ¨π‘’ , π‘£βŸ© = 0 ⟺ ‖𝑒‖ ≀ ‖𝑒 + π‘Žπ‘£β€– for all π‘Ž ∈ 𝐅 . 7 Suppose 𝑒 , 𝑣 ∈ 𝑉 . Prove that β€–π‘Žπ‘’ + 𝑏𝑣‖ = ‖𝑏𝑒 + π‘Žπ‘£β€– for all π‘Ž , 𝑏 ∈ 𝐑 if and only if ‖𝑒‖ = ‖𝑣‖ . 8 Suppose π‘Ž , 𝑏 , 𝑐 , π‘₯ , 𝑦 ∈ 𝐑 and π‘Ž 2 + 𝑏 2 + 𝑐 2 + π‘₯ 2 + 𝑦 2 ≀ 1 . Prove that π‘Ž + 𝑏 + 𝑐 + 4π‘₯ + 9𝑦 ≀ 10 . 9 Suppose 𝑒 , 𝑣 ∈ 𝑉 and ‖𝑒‖ = ‖𝑣‖ = 1 and βŸ¨π‘’ , π‘£βŸ© = 1 . Prove that 𝑒 = 𝑣 . 10 Suppose 𝑒 , 𝑣 ∈ 𝑉 and ‖𝑒‖ ≀ 1 and ‖𝑣‖ ≀ 1 . Prove that √ 1 βˆ’ ‖𝑒‖ 2 √ 1 βˆ’ ‖𝑣‖ 2 ≀ 1 βˆ’ βˆ£βŸ¨π‘’ , π‘£βŸ©βˆ£. 11 Find vectors 𝑒 , 𝑣 ∈ 𝐑 2 such that 𝑒 is a scalar multiple of (1 , 3) , 𝑣 is orthog- onal to (1 , 3) , and (1 , 2) = 𝑒 + 𝑣 . 12 Suppose π‘Ž , 𝑏 , 𝑐 , 𝑑 are positive numbers. (a) Prove that (π‘Ž + 𝑏 + 𝑐 + 𝑑)(1π‘Ž + 1 𝑏 + 1 𝑐 + 1 𝑑) β‰₯ 16 . (b) For which positive numbers π‘Ž , 𝑏 , 𝑐 , 𝑑 is the inequality above an equality? 13 Show that the square of an average is less than or equal to the average of the squares. More precisely, show that if π‘Ž 1 , … , π‘Ž 𝑛 ∈ 𝐑 , then the square of the average of π‘Ž 1 , … , π‘Ž 𝑛 is less than or equal to the average of π‘Ž 12 , … , π‘Ž 𝑛2 . 14 Suppose 𝑣 ∈ 𝑉 and 𝑣 β‰  0 . Prove that 𝑣/‖𝑣‖ is the unique closest element on the unit sphere of 𝑉 to 𝑣 . More precisely, prove that if 𝑒 ∈ 𝑉 and ‖𝑒‖ = 1 , then βˆ₯𝑣 βˆ’ 𝑣 ‖𝑣‖βˆ₯ ≀ ‖𝑣 βˆ’ 𝑒‖ , with equality only if 𝑒 = 𝑣/‖𝑣‖ . 15 Suppose 𝑒 , 𝑣 are nonzero vectors in 𝐑 2 . Prove that βŸ¨π‘’ , π‘£βŸ© = ‖𝑒‖ ‖𝑣‖ cos πœƒ , where πœƒ is the angle between 𝑒 and 𝑣 (thinking of 𝑒 and 𝑣 as arrows with initial point at the origin). Hint: Use the law of cosines on the triangle formed by 𝑒 , 𝑣 , and 𝑒 βˆ’ 𝑣 . Section 6A Inner Products and Norms 193 16 The angle between two vectors (thought of as arrows with initial point at the origin) in 𝐑 2 or 𝐑 3 can be defined geometrically. However, geometry is not as clear in 𝐑 𝑛 for 𝑛 > 3 . Thus the angle between two nonzero vectors π‘₯ , 𝑦 ∈ 𝐑 𝑛 is defined to be arccos ⟨π‘₯ , π‘¦βŸ© β€–π‘₯β€– ‖𝑦‖ , where the motivation for this definition comes from Exercise 15. Explain why the Cauchy–Schwarz inequality is needed to show that this definition makes sense. 17 Prove that ( 𝑛 βˆ‘ π‘˜=1 π‘Ž π‘˜ 𝑏 π‘˜ ) 2 ≀ ( 𝑛 βˆ‘ π‘˜=1 π‘˜π‘Ž π‘˜2 )( 𝑛 βˆ‘ π‘˜=1 𝑏 π‘˜2 π‘˜ ) for all real numbers π‘Ž 1 , … , π‘Ž 𝑛 and 𝑏 1 , … , 𝑏 𝑛 . 18 (a) Suppose 𝑓 ∢ [1 , ∞) β†’ [0 , ∞) is continuous. Show that ( ∫ ∞ 1 𝑓 ) 2 ≀ ∫ ∞ 1 π‘₯ 2 ( 𝑓 (π‘₯)) 2 𝑑π‘₯. (b) For which continuous functions 𝑓 ∢ [1 , ∞) β†’ [0 , ∞) is the inequality in (a) an equality with both sides finite? 19 Suppose 𝑣 1 , … , 𝑣 𝑛 is a basis of 𝑉 and 𝑇 ∈ β„’ (𝑉) . Prove that if πœ† is an eigenvalue of 𝑇 , then |πœ†| 2 ≀ 𝑛 βˆ‘ 𝑗 = 1 𝑛 βˆ‘ π‘˜=1 | β„³ (𝑇) 𝑗 , π‘˜ | 2 , where β„³ (𝑇) 𝑗 , π‘˜ denotes the entry in row 𝑗 , column π‘˜ of the matrix of 𝑇 with respect to the basis 𝑣 1 , … , 𝑣 𝑛 . 20 Prove that if 𝑒 , 𝑣 ∈ 𝑉 , then ∣ ‖𝑒‖ βˆ’ ‖𝑣‖ ∣ ≀ ‖𝑒 βˆ’ 𝑣‖ . The inequality above is called the reverse triangle inequality . For the reverse triangle inequality when 𝑉 = 𝐂 , see Exercise 2 in Chapter 4. 21 Suppose 𝑒 , 𝑣 ∈ 𝑉 are such that ‖𝑒‖ = 3 , ‖𝑒 + 𝑣‖ = 4 , ‖𝑒 βˆ’ 𝑣‖ = 6. What number does ‖𝑣‖ equal? 22 Show that if 𝑒 , 𝑣 ∈ 𝑉 , then ‖𝑒 + 𝑣‖ ‖𝑒 βˆ’ 𝑣‖ ≀ ‖𝑒‖ 2 + ‖𝑣‖ 2 . 23 Suppose 𝑣 1 , … , 𝑣 π‘š ∈ 𝑉 are such that ‖𝑣 π‘˜ β€– ≀ 1 for each π‘˜ = 1 , … , π‘š . Show that there exist π‘Ž 1 , … , π‘Ž π‘š ∈ {1 , βˆ’1} such that β€–π‘Ž 1 𝑣 1 + β‹― + π‘Ž π‘š 𝑣 π‘š β€– ≀ √ π‘š. 194 Chapter 6 Inner Product Spaces 24 Prove or give a counterexample: If β€–β‹…β€– is the norm associated with an inner product on 𝐑 2 , then there exists (π‘₯ , 𝑦) ∈ 𝐑 2 such that β€–(π‘₯ , 𝑦)β€– β‰  max {|π‘₯| , |𝑦|} . 25 Suppose 𝑝 > 0 . Prove that there is an inner product on 𝐑 2 such that the associated norm is given by β€–(π‘₯ , 𝑦)β€– = (|π‘₯| 𝑝 + |𝑦| 𝑝 ) 1/𝑝 for all (π‘₯ , 𝑦) ∈ 𝐑 2 if and only if 𝑝 = 2 . 26 Suppose 𝑉 is a real inner product space. Prove that βŸ¨π‘’ , π‘£βŸ© = ‖𝑒 + 𝑣‖ 2 βˆ’ ‖𝑒 βˆ’ 𝑣‖ 2 4 for all 𝑒 , 𝑣 ∈ 𝑉 . 27 Suppose 𝑉 is a complex inner product space. Prove that βŸ¨π‘’ , π‘£βŸ© = ‖𝑒 + 𝑣‖ 2 βˆ’ ‖𝑒 βˆ’ 𝑣‖ 2 + ‖𝑒 + 𝑖𝑣‖ 2 𝑖 βˆ’ ‖𝑒 βˆ’ 𝑖𝑣‖ 2 𝑖 4 for all 𝑒 , 𝑣 ∈ 𝑉 . 28 A norm on a vector space π‘ˆ is a function β€–β‹…β€– ∢ π‘ˆ β†’ [0 , ∞) such that ‖𝑒‖ = 0 if and only if 𝑒 = 0 , ‖𝛼𝑒‖ = |𝛼|‖𝑒‖ for all 𝛼 ∈ 𝐅 and all 𝑒 ∈ π‘ˆ , and ‖𝑒 + 𝑣‖ ≀ ‖𝑒‖ + ‖𝑣‖ for all 𝑒 , 𝑣 ∈ π‘ˆ . Prove that a norm satisfying the parallelogram equality comes from an inner product ( in other words, show that if β€–β‹…β€– is a norm on π‘ˆ satisfying the parallelogram equality, then there is an inner product βŸ¨β‹… , β‹…βŸ© on π‘ˆ such that ‖𝑒‖ = βŸ¨π‘’ , π‘’βŸ© 1/2 for all 𝑒 ∈ π‘ˆ) . 29 Suppose 𝑉 1 , … , 𝑉 π‘š are inner product spaces. Show that the equation ⟨(𝑒 1 , … , 𝑒 π‘š ) , (𝑣 1 , … , 𝑣 π‘š )⟩ = βŸ¨π‘’ 1 , 𝑣 1 ⟩ + β‹― + βŸ¨π‘’ π‘š , 𝑣 π‘š ⟩ defines an inner product on 𝑉 1 Γ— β‹― Γ— 𝑉 π‘š . In the expression above on the right, for each π‘˜ = 1 , … , π‘š , the inner product βŸ¨π‘’ π‘˜ , 𝑣 π‘˜ ⟩ denotes the inner product on 𝑉 π‘˜ . Each of the spaces 𝑉 1 , … , 𝑉 π‘š may have a different inner product, even though the same notation is used here. 30 Suppose 𝑉 is a real inner product space. For 𝑒 , 𝑣 , 𝑀 , π‘₯ ∈ 𝑉 , define βŸ¨π‘’ + 𝑖𝑣 , 𝑀 + 𝑖π‘₯⟩ 𝐂 = βŸ¨π‘’ , π‘€βŸ© + βŸ¨π‘£ , π‘₯⟩ + (βŸ¨π‘£ , π‘€βŸ© βˆ’ βŸ¨π‘’ , π‘₯⟩)𝑖. (a) Show that βŸ¨β‹… , β‹…βŸ© 𝐂 makes 𝑉 𝐂 into a complex inner product space. (b) Show that if 𝑒 , 𝑣 ∈ 𝑉 , then βŸ¨π‘’ , π‘£βŸ© 𝐂 = βŸ¨π‘’ , π‘£βŸ© and ‖𝑒 + 𝑖𝑣‖ 𝐂2 = ‖𝑒‖ 2 + ‖𝑣‖ 2 . See Exercise 8 in Section 1B for the definition of the complexification 𝑉 𝐂 . Section 6A Inner Products and Norms 195 31 Suppose 𝑒 , 𝑣 , 𝑀 ∈ 𝑉 . Prove that βˆ₯𝑀 βˆ’ 12 (𝑒 + 𝑣)βˆ₯ 2 = ‖𝑀 βˆ’ 𝑒‖ 2 + ‖𝑀 βˆ’ 𝑣‖ 2 2 βˆ’ ‖𝑒 βˆ’ 𝑣‖ 2 4 . 32 Suppose that 𝐸 is a subset of 𝑉 with the property that 𝑒 , 𝑣 ∈ 𝐸 implies 12 (𝑒 + 𝑣) ∈ 𝐸 . Let 𝑀 ∈ 𝑉 . Show that there is at most one point in 𝐸 that is closest to 𝑀 . In other words, show that there is at most one 𝑒 ∈ 𝐸 such that ‖𝑀 βˆ’ 𝑒‖ ≀ ‖𝑀 βˆ’ π‘₯β€– for all π‘₯ ∈ 𝐸 . 33 Suppose 𝑓 , 𝑔 are differentiable functions from 𝐑 to 𝐑 𝑛 . (a) Show that ⟨ 𝑓 (𝑑) , 𝑔(𝑑)⟩ β€² = ⟨ 𝑓 β€² (𝑑) , 𝑔(𝑑)⟩ + ⟨ 𝑓 (𝑑) , 𝑔 β€² (𝑑)⟩. (b) Suppose 𝑐 is a positive number and βˆ₯ 𝑓 (𝑑)βˆ₯ = 𝑐 for every 𝑑 ∈ 𝐑 . Show that ⟨ 𝑓 β€² (𝑑) , 𝑓 (𝑑)⟩ = 0 for every 𝑑 ∈ 𝐑 . (c) Interpret the result in (b) geometrically in terms of the tangent vector to a curve lying on a sphere in 𝐑 𝑛 centered at the origin. A function 𝑓 ∢ 𝐑 β†’ 𝐑 𝑛 is called differentiable if there exist differentiable functions 𝑓 1 , … , 𝑓 𝑛 from 𝐑 to 𝐑 such that 𝑓(𝑑) = ( 𝑓 1 (𝑑) , … , 𝑓 𝑛 (𝑑)) for each 𝑑 ∈ 𝐑 . Furthermore, for each 𝑑 ∈ 𝐑 , the derivative 𝑓 β€² (𝑑) ∈ 𝐑 𝑛 is defined by 𝑓 β€² (𝑑) = ( 𝑓 1β€² (𝑑) , … , 𝑓 𝑛′ (𝑑)) . 34 Use inner products to prove Apollonius’s identity: In a triangle with sides of length π‘Ž , 𝑏 , and 𝑐 , let 𝑑 be the length of the line segment from the midpoint of the side of length 𝑐 to the opposite vertex. Then π‘Ž 2 + 𝑏 2 = 1 2 𝑐 2 + 2𝑑 2 . Annotated Entity: ID: 209 Spans: True Boxes: True Text: 196 Chapter 6 Inner Product Spaces 35 Fix a positive integer 𝑛 . The Laplacian Δ𝑝 of a twice differentiable real- valued function 𝑝 on 𝐑 𝑛 is the function on 𝐑 𝑛 defined by Δ𝑝 = πœ• 2 𝑝 πœ•π‘₯ 12 + β‹― + πœ• 2 𝑝 πœ•π‘₯ 𝑛2 . The function 𝑝 is called harmonic if Δ𝑝 = 0 . A polynomial on 𝐑 𝑛 is a linear combination (with coefficients in 𝐑 ) of functions of the form π‘₯ 1π‘š 1 β‹―π‘₯ π‘›π‘š 𝑛 , where π‘š 1 , … , π‘š 𝑛 are nonnegative integers. Suppose π‘ž is a polynomial on 𝐑 𝑛 . Prove that there exists a harmonic polynomial 𝑝 on 𝐑 𝑛 such that 𝑝(π‘₯) = π‘ž(π‘₯) for every π‘₯ ∈ 𝐑 𝑛 with β€–π‘₯β€– = 1 . The only fact about harmonic functions that you need for this exercise is that if 𝑝 is a harmonic function on 𝐑 𝑛 and 𝑝(π‘₯) = 0 for all π‘₯ ∈ 𝐑 𝑛 with β€–π‘₯β€– = 1 , then 𝑝 = 0 . Hint: A reasonable guess is that the desired harmonic polynomial 𝑝 is of the form π‘ž + (1βˆ’β€–π‘₯β€– 2 )π‘Ÿ for some polynomial π‘Ÿ . Prove that there is a polynomial π‘Ÿ on 𝐑 𝑛 such that π‘ž + (1 βˆ’ β€–π‘₯β€– 2 )π‘Ÿ is harmonic by defining an operator 𝑇 on a suitable vector space by π‘‡π‘Ÿ = Ξ”((1 βˆ’ β€–π‘₯β€– 2 )π‘Ÿ) and then showing that 𝑇 is injective and hence surjective. In realms of numbers, where the secrets lie, A noble truth emerges from the deep, Cauchy and Schwarz, their wisdom they apply, An inequality for all to keep. Two vectors, by this bond, are intertwined, As inner products weave a gilded thread, Their magnitude, by providence, confined, A bound to which their destiny is wed. Though shadows fall, and twilight dims the day, This inequality will stand the test, To guide us in our quest, to light the way, And in its truth, our understanding rest. So sing, ye muses, of this noble feat, Cauchy–Schwarz, the bound that none can beat. β€”written by ChatGPT with input Shakespearean sonnet on Cauchy–Schwarz inequality Annotated Entity: ID: 210 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 197 6B Orthonormal Bases Orthonormal Lists and the Gram–Schmidt Procedure 6.22 definition: orthonormal β€’ A list of vectors is called orthonormal if each vector in the list has norm 1 and is orthogonal to all the other vectors in the list. β€’ In other words, a list 𝑒 1 , … , 𝑒 π‘š of vectors in 𝑉 is orthonormal if βŸ¨π‘’ 𝑗 , 𝑒 π‘˜ ⟩ = ⎧{ ⎨{⎩ 1 if 𝑗 = π‘˜ , 0 if 𝑗 β‰  π‘˜ for all 𝑗 , π‘˜ ∈ {1 , … , π‘š} . 6.23 example: orthonormal lists (a) The standard basis of 𝐅 𝑛 is an orthonormal list. (b) ( 1 √ 3 , 1 √ 3 , 1 √ 3 ) , (βˆ’ 1 √ 2 , 1 √ 2 , 0) is an orthonormal list in 𝐅 3 . (c) ( 1 √ 3 , 1 √ 3 , 1 √ 3 ) , (βˆ’ 1 √ 2 , 1 √ 2 , 0) , ( 1 √ 6 , 1 √ 6 , βˆ’ 2 √ 6 ) is an orthonormal list in 𝐅 3 . (d) Suppose 𝑛 is a positive integer. Then, as Exercise 4 asks you to verify, 1 √ 2πœ‹ , cos π‘₯ √ πœ‹ , cos 2π‘₯ √ πœ‹ , … , cos 𝑛π‘₯ √ πœ‹ , sin π‘₯ √ πœ‹ , sin 2π‘₯ √ πœ‹ , … , sin 𝑛π‘₯ √ πœ‹ is an orthonormal list of vectors in 𝐢[βˆ’πœ‹ , πœ‹] , the vector space of continuous real-valued functions on [βˆ’πœ‹ , πœ‹] with inner product ⟨ 𝑓 , π‘”βŸ© = ∫ πœ‹ βˆ’πœ‹ 𝑓 𝑔. The orthonormal list above is often used for modeling periodic phenomena, such as tides. (e) Suppose we make 𝒫 2 (𝐑) into an inner product space using the inner product given by βŸ¨π‘ , π‘žβŸ© = ∫ 1 βˆ’1 π‘π‘ž for all 𝑝 , π‘ž ∈ 𝒫 2 (𝐑) . The standard basis 1 , π‘₯ , π‘₯ 2 of 𝒫 2 (𝐑) is not an orthonor- mal list because the vectors in that list do not have norm 1 . Dividing each vector by its norm gives the list 1/ √ 2 , √ 3/2π‘₯ , √ 5/2π‘₯ 2 , in which each vector has norm 1 , and the second vector is orthogonal to the first and third vectors. However, the first and third vectors are not orthogonal. Thus this is not an orthonormal list. Soon we will see how to construct an orthonormal list from the standard basis 1 , π‘₯ , π‘₯ 2 (see Example 6.34). Annotated Entity: ID: 211 Spans: True Boxes: True Text: 198 Chapter 6 Inner Product Spaces Orthonormal lists are particularly easy to work with, as illustrated by the next result. 6.24 norm of an orthonormal linear combination Suppose 𝑒 1 , … , 𝑒 π‘š is an orthonormal list of vectors in 𝑉 . Then β€–π‘Ž 1 𝑒 1 + β‹― + π‘Ž π‘š 𝑒 π‘š β€– 2 = |π‘Ž 1 | 2 + β‹― + |π‘Ž π‘š | 2 for all π‘Ž 1 , … , π‘Ž π‘š ∈ 𝐅 . Proof Because each 𝑒 π‘˜ has norm 1 , this follows from repeated applications of the Pythagorean theorem (6.12). The result above has the following important corollary. 6.25 orthonormal lists are linearly independent Every orthonormal list of vectors is linearly independent. Proof Suppose 𝑒 1 , … , 𝑒 π‘š is an orthonormal list of vectors in 𝑉 and π‘Ž 1 , … , π‘Ž π‘š ∈ 𝐅 are such that π‘Ž 1 𝑒 1 + β‹― + π‘Ž π‘š 𝑒 π‘š = 0. Then |π‘Ž 1 | 2 + β‹― + |π‘Ž π‘š | 2 = 0 (by 6.24), which means that all the π‘Ž π‘˜ ’s are 0 . Thus 𝑒 1 , … , 𝑒 π‘š is linearly independent. Now we come to an important inequality. 6.26 Bessel’s inequality Suppose 𝑒 1 , … , 𝑒 π‘š is an orthonormal list of vectors in 𝑉 . If 𝑣 ∈ 𝑉 then βˆ£βŸ¨π‘£ , 𝑒 1 ⟩∣ 2 + β‹― + βˆ£βŸ¨π‘£ , 𝑒 π‘š ⟩∣ 2 ≀ ‖𝑣‖ 2 . Proof Suppose 𝑣 ∈ 𝑉 . Then 𝑣 = βŸ¨π‘£ , 𝑒 1 βŸ©π‘’ 1 + β‹― + βŸ¨π‘£ , 𝑒 π‘š βŸ©π‘’ π‘š ⏟⏟⏟⏟⏟⏟⏟⏟⏟ 𝑒 + 𝑣 βˆ’ βŸ¨π‘£ , 𝑒 1 βŸ©π‘’ 1 βˆ’ β‹― βˆ’ βŸ¨π‘£ , 𝑒 π‘š βŸ©π‘’ π‘š ⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟ 𝑀 . Let 𝑒 and 𝑀 be defined as in the equation above. If π‘˜ ∈ {1 , … , π‘š} , then βŸ¨π‘€ , 𝑒 π‘˜ ⟩ = βŸ¨π‘£ , 𝑒 π‘˜ ⟩ βˆ’ βŸ¨π‘£ , 𝑒 π‘˜ βŸ©βŸ¨π‘’ π‘˜ , 𝑒 π‘˜ ⟩ = 0 . This implies that βŸ¨π‘€ , π‘’βŸ© = 0 . The Pythagorean theorem now implies that ‖𝑣‖ 2 = ‖𝑒‖ 2 + ‖𝑀‖ 2 β‰₯ ‖𝑒‖ 2 = βˆ£βŸ¨π‘£ , 𝑒 1 ⟩∣ 2 + β‹― + βˆ£βŸ¨π‘£ , 𝑒 π‘š ⟩∣ 2 , where the last line comes from 6.24. Annotated Entity: ID: 212 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 199 The next definition introduces one of the most useful concepts in the study of inner product spaces. 6.27 definition: orthonormal basis An orthonormal basis of 𝑉 is an orthonormal list of vectors in 𝑉 that is also a basis of 𝑉 . For example, the standard basis is an orthonormal basis of 𝐅 𝑛 . 6.28 orthonormal lists of the right length are orthonormal bases Suppose 𝑉 is finite-dimensional. Then every orthonormal list of vectors in 𝑉 of length dim 𝑉 is an orthonormal basis of 𝑉 . Proof By 6.25, every orthonormal list of vectors in 𝑉 is linearly independent. Thus every such list of the right length is a basisβ€”see 2.38. 6.29 example: an orthonormal basis of 𝐅 4 As mentioned above, the standard basis is an orthonormal basis of 𝐅 4 . We now show that ( 12 , 12 , 12 , 12 ) , ( 12 , 12 , βˆ’ 12 , βˆ’ 12 ) , ( 12 , βˆ’ 12 , βˆ’ 12 , 12 ) , (βˆ’ 12 , 12 , βˆ’ 12 , 12 ) is also an orthonormal basis of 𝐅 4 . We have βˆ₯( 1 2 , 1 2 , 1 2 , 1 2 )βˆ₯ = √ 1 2 2 + 1 2 2 + 1 2 2 + 1 2 2 = 1. Similarly, the other three vectors in the list above also have norm 1 . Note that ⟨( 12 , 12 , 12 , 12 ) , ( 12 , 12 , βˆ’ 12 , βˆ’ 12 )⟩ = 12 β‹… 12 + 12 β‹… 12 + 12 β‹… (βˆ’ 12 ) + 12 β‹… (βˆ’ 12 ) = 0. Similarly, the inner product of any two distinct vectors in the list above also equals 0 . Thus the list above is orthonormal. Because we have an orthonormal list of length four in the four-dimensional vector space 𝐅 4 , this list is an orthonormal basis of 𝐅 4 (by 6.28). In general, given a basis 𝑒 1 , … , 𝑒 𝑛 of 𝑉 and a vector 𝑣 ∈ 𝑉 , we know that there is some choice of scalars π‘Ž 1 , … , π‘Ž 𝑛 ∈ 𝐅 such that 𝑣 = π‘Ž 1 𝑒 1 + β‹― + π‘Ž 𝑛 𝑒 𝑛 . Computing the numbers π‘Ž 1 , … , π‘Ž 𝑛 that satisfy the equation above can be a long computation for an arbitrary basis of 𝑉 . The next result shows, however, that this is easy for an orthonormal basisβ€”just take π‘Ž π‘˜ = βŸ¨π‘£ , 𝑒 π‘˜ ⟩ . Annotated Entity: ID: 213 Spans: True Boxes: True Text: 200 Chapter 6 Inner Product Spaces The formula below for ‖𝑣‖ is called Parseval’s identity. It was published in 1799 in the context of Fourier series. Notice how the next result makes each inner product space of dimension 𝑛 behave like 𝐅 𝑛 , with the role of the coordinates of a vector in 𝐅 𝑛 played by βŸ¨π‘£ , 𝑒 1 ⟩ , … , βŸ¨π‘£ , 𝑒 𝑛 ⟩ . 6.30 writing a vector as a linear combination of an orthonormal basis Suppose 𝑒 1 , … , 𝑒 𝑛 is an orthonormal basis of 𝑉 and 𝑒 , 𝑣 ∈ 𝑉 . Then (a) 𝑣 = βŸ¨π‘£ , 𝑒 1 βŸ©π‘’ 1 + β‹― + βŸ¨π‘£ , 𝑒 𝑛 βŸ©π‘’ 𝑛 , (b) ‖𝑣‖ 2 = βˆ£βŸ¨π‘£ , 𝑒 1 ⟩∣ 2 + β‹― + βˆ£βŸ¨π‘£ , 𝑒 𝑛 ⟩∣ 2 , (c) βŸ¨π‘’ , π‘£βŸ© = βŸ¨π‘’ , 𝑒 1 βŸ©βŸ¨π‘£ , 𝑒 1 ⟩ + β‹― + βŸ¨π‘’ , 𝑒 𝑛 βŸ©βŸ¨π‘£ , 𝑒 𝑛 ⟩ . Proof Because 𝑒 1 , … , 𝑒 𝑛 is a basis of 𝑉 , there exist scalars π‘Ž 1 , … , π‘Ž 𝑛 such that 𝑣 = π‘Ž 1 𝑒 1 + β‹― + π‘Ž 𝑛 𝑒 𝑛 . Because 𝑒 1 , … , 𝑒 𝑛 is orthonormal, taking the inner product of both sides of this equation with 𝑒 π‘˜ gives βŸ¨π‘£ , 𝑒 π‘˜ ⟩ = π‘Ž π‘˜ . Thus (a) holds. Now (b) follows immediately from (a) and 6.24. Take the inner product of 𝑒 with each side of (a) and then get (c) by using conjugate linearity [6.6(d) and 6.6(e)] in the second slot of the inner product. 6.31 example: finding coefficients for a linear combination Suppose we want to write the vector (1 , 2 , 4 , 7) ∈ 𝐅 4 as a linear combination of the orthonormal basis ( 12 , 12 , 12 , 12 ) , ( 12 , 12 , βˆ’ 12 , βˆ’ 12 ) , ( 12 , βˆ’ 12 , βˆ’ 12 , 12 ) , (βˆ’ 12 , 12 , βˆ’ 12 , 12 ) of 𝐅 4 from Example 6.29. Instead of solving a system of four linear equations in four unknowns, as typically would be required if we were working with a nonorthonormal basis, we simply evaluate four inner products and use 6.30(a), getting that (1 , 2 , 4 , 7) equals 7( 12 , 12 , 12 , 12 ) βˆ’ 4( 12 , 12 , βˆ’ 12 , βˆ’ 12 ) + ( 12 , βˆ’ 12 , βˆ’ 12 , 12 ) + 2(βˆ’ 12 , 12 , βˆ’ 12 , 12 ). Now that we understand the usefulness of orthonormal bases, how do we go about finding them? For example, does 𝒫 π‘š (𝐑) with inner product as in 6.3(c) have an orthonormal basis? The next result will lead to answers to these questions. JΓΈrgen Gram ( 1850–1916 ) and Erhard Schmidt ( 1876–1959 ) popularized this algorithm that constructs orthonormal lists. The algorithm used in the next proof is called the Gram–Schmidt procedure . It gives a method for turning a linearly independent list into an orthonormal list with the same span as the original list. Annotated Entity: ID: 214 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 201 6.32 Gram–Schmidt procedure Suppose 𝑣 1 , … , 𝑣 π‘š is a linearly independent list of vectors in 𝑉 . Let 𝑓 1 = 𝑣 1 . For π‘˜ = 2 , … , π‘š , define 𝑓 π‘˜ inductively by 𝑓 π‘˜ = 𝑣 π‘˜ βˆ’ βŸ¨π‘£ π‘˜ , 𝑓 1 ⟩ β€– 𝑓 1 β€– 2 𝑓 1 βˆ’ β‹― βˆ’ βŸ¨π‘£ π‘˜ , 𝑓 π‘˜βˆ’1 ⟩ β€– 𝑓 π‘˜βˆ’1 β€– 2 𝑓 π‘˜βˆ’1 . For each π‘˜ = 1 , … , π‘š , let 𝑒 π‘˜ = 𝑓 π‘˜ β€– 𝑓 π‘˜ β€– . Then 𝑒 1 , … , 𝑒 π‘š is an orthonormal list of vectors in 𝑉 such that span (𝑣 1 , … , 𝑣 π‘˜ ) = span (𝑒 1 , … , 𝑒 π‘˜ ) for each π‘˜ = 1 , … , π‘š . Proof We will show by induction on π‘˜ that the desired conclusion holds. To get started with π‘˜ = 1 , note that because 𝑒 1 = 𝑓 1 /β€– 𝑓 1 β€– , we have ‖𝑒 1 β€– = 1 ; also, span (𝑣 1 ) = span (𝑒 1 ) because 𝑒 1 is a nonzero multiple of 𝑣 1 . Suppose 1 < π‘˜ ≀ π‘š and the list 𝑒 1 , … , 𝑒 π‘˜βˆ’1 generated by 6.32 is an orthonormal list such that 6.33 span (𝑣 1 , … , 𝑣 π‘˜βˆ’1 ) = span (𝑒 1 , … , 𝑒 π‘˜βˆ’1 ). Because 𝑣 1 , … , 𝑣 π‘š is linearly independent, we have 𝑣 π‘˜ βˆ‰ span (𝑣 1 , … , 𝑣 π‘˜βˆ’1 ) . Thus 𝑣 π‘˜ βˆ‰ span (𝑒 1 , … , 𝑒 π‘˜βˆ’1 ) = span ( 𝑓 1 , … , 𝑓 π‘˜βˆ’1 ) , which implies that 𝑓 π‘˜ β‰  0 . Hence we are not dividing by 0 in the definition of 𝑒 π‘˜ given in 6.32. Dividing a vector by its norm produces a new vector with norm 1 ; thus ‖𝑒 π‘˜ β€– = 1 . Let 𝑗 ∈ {1 , … , π‘˜ βˆ’ 1} . Then βŸ¨π‘’ π‘˜ , 𝑒 𝑗 ⟩ = 1 β€– 𝑓 π‘˜ β€– β€– 𝑓 𝑗 β€–βŸ¨ 𝑓 π‘˜ , 𝑓 𝑗 ⟩ = 1 β€– 𝑓 π‘˜ β€– β€– 𝑓 𝑗 β€–βŸ¨π‘£ π‘˜ βˆ’ βŸ¨π‘£ π‘˜ , 𝑓 1 ⟩ β€– 𝑓 1 β€– 2 𝑓 1 βˆ’ β‹― βˆ’ βŸ¨π‘£ π‘˜ , 𝑓 π‘˜βˆ’1 ⟩ β€– 𝑓 π‘˜βˆ’1 β€– 2 𝑓 π‘˜βˆ’1 , 𝑓 𝑗 ⟩ = 1 β€– 𝑓 π‘˜ β€– β€– 𝑓 𝑗 β€–(βŸ¨π‘£ π‘˜ , 𝑓 𝑗 ⟩ βˆ’ βŸ¨π‘£ π‘˜ , 𝑓 𝑗 ⟩) = 0. Thus 𝑒 1 , … , 𝑒 π‘˜ is an orthonormal list. From the definition of 𝑒 π‘˜ given in 6.32, we see that 𝑣 π‘˜ ∈ span (𝑒 1 , … , 𝑒 π‘˜ ) . Combining this information with 6.33 shows that span (𝑣 1 , … , 𝑣 π‘˜ ) βŠ† span (𝑒 1 , … , 𝑒 π‘˜ ). Both lists above are linearly independent (the 𝑣 ’s by hypothesis, and the 𝑒 ’s by orthonormality and 6.25). Thus both subspaces above have dimension π‘˜ , and hence they are equal, completing the induction step and thus completing the proof. Annotated Entity: ID: 215 Spans: True Boxes: True Text: 202 Chapter 6 Inner Product Spaces 6.34 example: an orthonormal basis of 𝒫 2 (𝐑) Suppose we make 𝒫 2 (𝐑) into an inner product space using the inner product given by βŸ¨π‘ , π‘žβŸ© = ∫ 1 βˆ’1 π‘π‘ž for all 𝑝 , π‘ž ∈ 𝒫 2 (𝐑) . We know that 1 , π‘₯ , π‘₯ 2 is a basis of 𝒫 2 (𝐑) , but it is not an orthonormal basis. We will find an orthonormal basis of 𝒫 2 (𝐑) by applying the Gram–Schmidt procedure with 𝑣 1 = 1 , 𝑣 2 = π‘₯ , and 𝑣 3 = π‘₯ 2 . To get started, take 𝑓 1 = 𝑣 1 = 1 . Thus β€– 𝑓 1 β€– 2 = ∫ 1βˆ’1 1 = 2 . Hence the formula in 6.32 tells us that 𝑓 2 = 𝑣 2 βˆ’ βŸ¨π‘£ 2 , 𝑓 1 ⟩ β€– 𝑓 1 β€– 2 𝑓 1 = π‘₯ βˆ’ ⟨π‘₯ , 1⟩ β€– 𝑓 1 β€– 2 = π‘₯ , where the last equality holds because ⟨π‘₯ , 1⟩ = ∫ 1βˆ’1 𝑑 𝑑𝑑 = 0 . The formula above for 𝑓 2 implies that β€– 𝑓 2 β€– 2 = ∫ 1βˆ’1 𝑑 2 𝑑𝑑 = 23 . Now the formula in 6.32 tells us that 𝑓 3 = 𝑣 3 βˆ’ βŸ¨π‘£ 3 , 𝑓 1 ⟩ β€– 𝑓 1 β€– 2 𝑓 1 βˆ’ βŸ¨π‘£ 3 , 𝑓 2 ⟩ β€– 𝑓 2 β€– 2 𝑓 2 = π‘₯ 2 βˆ’ 12 ⟨π‘₯ 2 , 1⟩ βˆ’ 32 ⟨π‘₯ 2 , π‘₯⟩π‘₯ = π‘₯ 2 βˆ’ 13 . The formula above for 𝑓 3 implies that β€– 𝑓 3 β€– 2 = ∫ 1 βˆ’1 (𝑑 2 βˆ’ 13 ) 2 𝑑𝑑 = ∫ 1 βˆ’1 (𝑑 4 βˆ’ 23 𝑑 2 + 19 ) 𝑑𝑑 = 845 . Now dividing each of 𝑓 1 , 𝑓 2 , 𝑓 3 by its norm gives us the orthonormal list √ 12 , √ 32 π‘₯ , √ 458 (π‘₯ 2 βˆ’ 13 ). The orthonormal list above has length three, which is the dimension of 𝒫 2 (𝐑) . Hence this orthonormal list is an orthonormal basis of 𝒫 2 (𝐑) [by 6.28]. Now we can answer the question about the existence of orthonormal bases. 6.35 existence of orthonormal basis Every finite-dimensional inner product space has an orthonormal basis. Proof Suppose 𝑉 is finite-dimensional. Choose a basis of 𝑉 . Apply the Gram– Schmidt procedure (6.32) to it, producing an orthonormal list of length dim 𝑉 . By 6.28, this orthonormal list is an orthonormal basis of 𝑉 . Sometimes we need to know not only that an orthonormal basis exists, but also that every orthonormal list can be extended to an orthonormal basis. In the next corollary, the Gram–Schmidt procedure shows that such an extension is always possible. Annotated Entity: ID: 216 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 203 6.36 every orthonormal list extends to an orthonormal basis Suppose 𝑉 is finite-dimensional. Then every orthonormal list of vectors in 𝑉 can be extended to an orthonormal basis of 𝑉 . Proof Suppose 𝑒 1 , … , 𝑒 π‘š is an orthonormal list of vectors in 𝑉 . Then 𝑒 1 , … , 𝑒 π‘š is linearly independent (by 6.25). Hence this list can be extended to a basis 𝑒 1 , … , 𝑒 π‘š , 𝑣 1 , … , 𝑣 𝑛 of 𝑉 (see 2.32). Now apply the Gram–Schmidt procedure (6.32) to 𝑒 1 , … , 𝑒 π‘š , 𝑣 1 , … , 𝑣 𝑛 , producing an orthonormal list 𝑒 1 , … , 𝑒 π‘š , 𝑓 1 , … , 𝑓 𝑛 ; here the formula given by the Gram–Schmidt procedure leaves the first π‘š vectors unchanged because they are already orthonormal. The list above is an orthonormal basis of 𝑉 by 6.28. Recall that a matrix is called upper triangular if it looks like this: βŽ›βŽœβŽœβŽœβŽ βˆ— βˆ— β‹± 0 βˆ— ⎞⎟⎟⎟⎠ , where the 0 in the matrix above indicates that all entries below the diagonal equal 0 , and asterisks are used to denote entries on and above the diagonal. In the last chapter, we gave a necessary and sufficient condition for an operator to have an upper-triangular matrix with respect to some basis (see 5.44). Now that we are dealing with inner product spaces, we would like to know whether there exists an orthonormal basis with respect to which we have an upper-triangular matrix. The next result shows that the condition for an operator to have an upper- triangular matrix with respect to some orthonormal basis is the same as the condition to have an upper-triangular matrix with respect to an arbitrary basis. 6.37 upper-triangular matrix with respect to some orthonormal basis Suppose 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉) . Then 𝑇 has an upper- triangular matrix with respect to some orthonormal basis of 𝑉 if and only if the minimal polynomial of 𝑇 equals (𝑧 βˆ’ πœ† 1 )β‹―(𝑧 βˆ’ πœ† π‘š ) for some πœ† 1 , … , πœ† π‘š ∈ 𝐅 . Proof Suppose 𝑇 has an upper-triangular matrix with respect to some basis 𝑣 1 , … , 𝑣 𝑛 of 𝑉 . Thus span (𝑣 1 , … , 𝑣 π‘˜ ) is invariant under 𝑇 for each π‘˜ = 1 , … , 𝑛 (see 5.39). Apply the Gram–Schmidt procedure to 𝑣 1 , … , 𝑣 𝑛 , producing an orthonormal basis 𝑒 1 , … , 𝑒 𝑛 of 𝑉 . Because span (𝑒 1 , … , 𝑒 π‘˜ ) = span (𝑣 1 , … , 𝑣 π‘˜ ) for each π‘˜ (see 6.32), we conclude that span (𝑒 1 , … , 𝑒 π‘˜ ) is invariant under 𝑇 for each π‘˜ = 1 , … , 𝑛 . Thus, by 5.39, 𝑇 has an upper-triangular matrix with respect to the orthonormal basis 𝑒 1 , … , 𝑒 𝑛 . Now use 5.44 to complete the proof. Annotated Entity: ID: 217 Spans: True Boxes: True Text: 204 Chapter 6 Inner Product Spaces Issai Schur ( 1875–1941 ) published a proof of the next result in 1909. For complex vector spaces, the next result is an important application of the result above. See Exercise 20 for a ver- sion of Schur’s theorem that applies simultaneously to more than one operator. 6.38 Schur’s theorem Every operator on a finite-dimensional complex inner product space has an upper-triangular matrix with respect to some orthonormal basis. Proof The desired result follows from the second version of the fundamental theorem of algebra (4.13) and 6.37. Linear Functionals on Inner Product Spaces Because linear maps into the scalar field 𝐅 play a special role, we defined a special name for them and their vector space in Section 3F. Those definitions are repeated below in case you skipped Section 3F. 6.39 definition: linear functional, dual space, 𝑉 β€² β€’ A linear functional on 𝑉 is a linear map from 𝑉 to 𝐅 . β€’ The dual space of 𝑉 , denoted by 𝑉 β€² , is the vector space of all linear functionals on 𝑉 . In other words, 𝑉 β€² = β„’ (𝑉 , 𝐅) . 6.40 example: linear functional on 𝐅 3 The function πœ‘ ∢ 𝐅 3 β†’ 𝐅 defined by πœ‘(𝑧 1 , 𝑧 2 , 𝑧 3 ) = 2𝑧 1 βˆ’ 5𝑧 2 + 𝑧 3 is a linear functional on 𝐅 3 . We could write this linear functional in the form πœ‘(𝑧) = βŸ¨π‘§ , π‘€βŸ© for every 𝑧 ∈ 𝐅 3 , where 𝑀 = (2 , βˆ’5 , 1) . 6.41 example: linear functional on 𝒫 5 (𝐑) The function πœ‘ ∢ 𝒫 5 (𝐑) β†’ 𝐑 defined by πœ‘(𝑝) = ∫ 1 βˆ’1 𝑝(𝑑)( cos (πœ‹π‘‘)) 𝑑𝑑 is a linear functional on 𝒫 5 (𝐑) . Annotated Entity: ID: 218 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 205 The next result is named in honor of Frigyes Riesz ( 1880–1956 ) , who proved several theorems early in the twentieth century that look very much like the result below. If 𝑣 ∈ 𝑉 , then the map that sends 𝑒 to βŸ¨π‘’ , π‘£βŸ© is a linear functional on 𝑉 . The next result states that every linear func- tional on 𝑉 is of this form. For example, we can take 𝑣 = (2 , βˆ’5 , 1) in Example 6.40. Suppose we make the vector space 𝒫 5 (𝐑) into an inner product space by defining βŸ¨π‘ , π‘žβŸ© = ∫ 1βˆ’1 π‘π‘ž . Let πœ‘ be as in Example 6.41. It is not obvious that there exists π‘ž ∈ 𝒫 5 (𝐑) such that ∫ 1 βˆ’1 𝑝(𝑑)( cos (πœ‹π‘‘)) 𝑑𝑑 = βŸ¨π‘ , π‘žβŸ© for every 𝑝 ∈ 𝒫 5 (𝐑) [ we cannot take π‘ž(𝑑) = cos (πœ‹π‘‘) because that choice of π‘ž is not an element of 𝒫 5 (𝐑)] . The next result tells us the somewhat surprising result that there indeed exists a polynomial π‘ž ∈ 𝒫 5 (𝐑) such that the equation above holds for all 𝑝 ∈ 𝒫 5 (𝐑) . 6.42 Riesz representation theorem Suppose 𝑉 is finite-dimensional and πœ‘ is a linear functional on 𝑉 . Then there is a unique vector 𝑣 ∈ 𝑉 such that πœ‘(𝑒) = βŸ¨π‘’ , π‘£βŸ© for every 𝑒 ∈ 𝑉 . Proof First we show that there exists a vector 𝑣 ∈ 𝑉 such that πœ‘(𝑒) = βŸ¨π‘’ , π‘£βŸ© for every 𝑒 ∈ 𝑉 . Let 𝑒 1 , … , 𝑒 𝑛 be an orthonormal basis of 𝑉 . Then πœ‘(𝑒) = πœ‘(βŸ¨π‘’ , 𝑒 1 βŸ©π‘’ 1 + β‹― + βŸ¨π‘’ , 𝑒 𝑛 βŸ©π‘’ 𝑛 ) = βŸ¨π‘’ , 𝑒 1 βŸ©πœ‘(𝑒 1 ) + β‹― + βŸ¨π‘’ , 𝑒 𝑛 βŸ©πœ‘(𝑒 𝑛 ) = βŸ¨π‘’ , πœ‘(𝑒 1 )𝑒 1 + β‹― + πœ‘(𝑒 𝑛 )𝑒 𝑛 ⟩ for every 𝑒 ∈ 𝑉 , where the first equality comes from 6.30(a). Thus setting 6.43 𝑣 = πœ‘(𝑒 1 )𝑒 1 + β‹― + πœ‘(𝑒 𝑛 )𝑒 𝑛 , we have πœ‘(𝑒) = βŸ¨π‘’ , π‘£βŸ© for every 𝑒 ∈ 𝑉 , as desired. Now we prove that only one vector 𝑣 ∈ 𝑉 has the desired behavior. Suppose 𝑣 1 , 𝑣 2 ∈ 𝑉 are such that πœ‘(𝑒) = βŸ¨π‘’ , 𝑣 1 ⟩ = βŸ¨π‘’ , 𝑣 2 ⟩ for every 𝑒 ∈ 𝑉 . Then 0 = βŸ¨π‘’ , 𝑣 1 ⟩ βˆ’ βŸ¨π‘’ , 𝑣 2 ⟩ = βŸ¨π‘’ , 𝑣 1 βˆ’ 𝑣 2 ⟩ for every 𝑒 ∈ 𝑉 . Taking 𝑒 = 𝑣 1 βˆ’ 𝑣 2 shows that 𝑣 1 βˆ’ 𝑣 2 = 0 . Thus 𝑣 1 = 𝑣 2 , completing the proof of the uniqueness part of the result. Annotated Entity: ID: 219 Spans: True Boxes: True Text: 206 Chapter 6 Inner Product Spaces 6.44 example: computation illustrating Riesz representation theorem Suppose we want to find a polynomial π‘ž ∈ 𝒫 2 (𝐑) such that 6.45 ∫ 1 βˆ’1 𝑝(𝑑)( cos (πœ‹π‘‘)) 𝑑𝑑 = ∫ 1 βˆ’1 π‘π‘ž for every polynomial 𝑝 ∈ 𝒫 2 (𝐑) . To do this, we make 𝒫 2 (𝐑) into an inner product space by defining βŸ¨π‘ , π‘žβŸ© to be the right side of the equation above for 𝑝 , π‘ž ∈ 𝒫 2 (𝐑) . Note that the left side of the equation above does not equal the inner product in 𝒫 2 (𝐑) of 𝑝 and the function 𝑑 ↦ cos (πœ‹π‘‘) because this last function is not a polynomial. Define a linear functional πœ‘ on 𝒫 2 (𝐑) by letting πœ‘(𝑝) = ∫ 1 βˆ’1 𝑝(𝑑)( cos (πœ‹π‘‘)) 𝑑𝑑 for each 𝑝 ∈ 𝒫 2 (𝐑) . Now use the orthonormal basis from Example 6.34 and apply formula 6.43 from the proof of the Riesz representation theorem to see that if 𝑝 ∈ 𝒫 2 (𝐑) , then πœ‘(𝑝) = βŸ¨π‘ , π‘žβŸ© , where π‘ž(π‘₯) = ( ∫ 1 βˆ’1 √ 12 cos (πœ‹π‘‘) 𝑑𝑑) √ 12 + ( ∫ 1 βˆ’1 √ 32 𝑑 cos (πœ‹π‘‘) 𝑑𝑑) √ 32 π‘₯ + ( ∫ 1 βˆ’1 √ 458 (𝑑 2 βˆ’ 13 ) cos (πœ‹π‘‘) 𝑑𝑑) √ 458 (π‘₯ 2 βˆ’ 13 ). A bit of calculus applied to the equation above shows that π‘ž(π‘₯) = 15 2πœ‹ 2 (1 βˆ’ 3π‘₯ 2 ). The same procedure shows that if we want to find π‘ž ∈ 𝒫 5 (𝐑) such that 6.45 holds for all 𝑝 ∈ 𝒫 5 (𝐑) , then we should take π‘ž(π‘₯) = 105 8πœ‹ 4 ((27 βˆ’ 2πœ‹ 2 ) + (24πœ‹ 2 βˆ’ 270)π‘₯ 2 + (315 βˆ’ 30πœ‹ 2 )π‘₯ 4 ). Suppose 𝑉 is finite-dimensional and πœ‘ a linear functional on 𝑉 . Then 6.43 gives a formula for the vector 𝑣 that satisfies πœ‘(𝑒) = βŸ¨π‘’ , π‘£βŸ© for all 𝑒 ∈ 𝑉 . Specifically, we have 𝑣 = πœ‘(𝑒 1 )𝑒 1 + β‹― + πœ‘(𝑒 𝑛 )𝑒 𝑛 . The right side of the equation above seems to depend on the orthonormal basis 𝑒 1 , … , 𝑒 𝑛 as well as on πœ‘ . However, 6.42 tells us that 𝑣 is uniquely determined by πœ‘ . Thus the right side of the equation above is the same regardless of which orthonormal basis 𝑒 1 , … , 𝑒 𝑛 of 𝑉 is chosen. For two additional different proofs of the Riesz representation theorem, see 6.58 and also Exercise 13 in Section 6C. Annotated Entity: ID: 220 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 207 Exercises 6B 1 Suppose 𝑒 1 , … , 𝑒 π‘š is a list of vectors in 𝑉 such that β€–π‘Ž 1 𝑒 1 + β‹― + π‘Ž π‘š 𝑒 π‘š β€– 2 = |π‘Ž 1 | 2 + β‹― + |π‘Ž π‘š | 2 for all π‘Ž 1 , … , π‘Ž π‘š ∈ 𝐅 . Show that 𝑒 1 , … , 𝑒 π‘š is an orthonormal list. This exercise provides a converse to 6.24. 2 (a) Suppose πœƒ ∈ 𝐑 . Show that both ( cos πœƒ , sin πœƒ) , (βˆ’ sin πœƒ , cos πœƒ) and ( cos πœƒ , sin πœƒ) , ( sin πœƒ , βˆ’ cos πœƒ) are orthonormal bases of 𝐑 2 . (b) Show that each orthonormal basis of 𝐑 2 is of the form given by one of the two possibilities in (a). 3 Suppose 𝑒 1 , … , 𝑒 π‘š is an orthonormal list in 𝑉 and 𝑣 ∈ 𝑉 . Prove that ‖𝑣‖ 2 = βˆ£βŸ¨π‘£ , 𝑒 1 ⟩∣ 2 + β‹― + βˆ£βŸ¨π‘£ , 𝑒 π‘š ⟩∣ 2 ⟺ 𝑣 ∈ span (𝑒 1 , … , 𝑒 π‘š ). 4 Suppose 𝑛 is a positive integer. Prove that 1 √ 2πœ‹ , cos π‘₯ √ πœ‹ , cos 2π‘₯ √ πœ‹ , … , cos 𝑛π‘₯ √ πœ‹ , sin π‘₯ √ πœ‹ , sin 2π‘₯ √ πœ‹ , … , sin 𝑛π‘₯ √ πœ‹ is an orthonormal list of vectors in 𝐢[βˆ’πœ‹ , πœ‹] , the vector space of continuous real-valued functions on [βˆ’πœ‹ , πœ‹] with inner product ⟨ 𝑓 , π‘”βŸ© = ∫ πœ‹ βˆ’πœ‹ 𝑓 𝑔. Hint: The following formulas should help. ( sin π‘₯)( cos 𝑦) = sin (π‘₯ βˆ’ 𝑦) + sin (π‘₯ + 𝑦) 2 ( sin π‘₯)( sin 𝑦) = cos (π‘₯ βˆ’ 𝑦) βˆ’ cos (π‘₯ + 𝑦) 2 ( cos π‘₯)( cos 𝑦) = cos (π‘₯ βˆ’ 𝑦) + cos (π‘₯ + 𝑦) 2 5 Suppose 𝑓 ∢ [βˆ’πœ‹ , πœ‹] β†’ 𝐑 is continuous. For each nonnegative integer π‘˜ , define π‘Ž π‘˜ = 1 √ πœ‹ ∫ πœ‹ βˆ’πœ‹ 𝑓 (π‘₯) cos (π‘˜π‘₯) 𝑑π‘₯ and 𝑏 π‘˜ = 1 √ πœ‹ ∫ πœ‹ βˆ’πœ‹ 𝑓 (π‘₯) sin (π‘˜π‘₯) 𝑑π‘₯. Prove that π‘Ž 02 2 + ∞ βˆ‘ π‘˜=1 (π‘Ž π‘˜2 + 𝑏 π‘˜2 ) ≀ ∫ πœ‹ βˆ’πœ‹ 𝑓 2 . The inequality above is actually an equality for all continuous functions 𝑓 ∢ [βˆ’πœ‹ , πœ‹] β†’ 𝐑 . However, proving that this inequality is an equality involves Fourier series techniques beyond the scope of this book. Annotated Entity: ID: 221 Spans: True Boxes: True Text: 208 Chapter 6 Inner Product Spaces 6 Suppose 𝑒 1 , … , 𝑒 𝑛 is an orthonormal basis of 𝑉 . (a) Prove that if 𝑣 1 , … , 𝑣 𝑛 are vectors in 𝑉 such that ‖𝑒 π‘˜ βˆ’ 𝑣 π‘˜ β€– < 1 √ 𝑛 for each π‘˜ , then 𝑣 1 , … , 𝑣 𝑛 is a basis of 𝑉 . (b) Show that there exist 𝑣 1 , … , 𝑣 𝑛 ∈ 𝑉 such that ‖𝑒 π‘˜ βˆ’ 𝑣 π‘˜ β€– ≀ 1 √ 𝑛 for each π‘˜ , but 𝑣 1 , … , 𝑣 𝑛 is not linearly independent. This exercise states in ( a ) that an appropriately small perturbation of an orthonormal basis is a basis. Then ( b ) shows that the number 1/ √ 𝑛 on the right side of the inequality in ( a ) cannot be improved upon. 7 Suppose 𝑇 ∈ β„’ (𝐑 3 ) has an upper-triangular matrix with respect to the basis (1 , 0 , 0) , (1, 1, 1), (1 , 1 , 2) . Find an orthonormal basis of 𝐑 3 with respect to which 𝑇 has an upper-triangular matrix. 8 Make 𝒫 2 (𝐑) into an inner product space by defining βŸ¨π‘ , π‘žβŸ© = ∫ 10 π‘π‘ž for all 𝑝 , π‘ž ∈ 𝒫 2 (𝐑) . (a) Apply the Gram–Schmidt procedure to the basis 1 , π‘₯ , π‘₯ 2 to produce an orthonormal basis of 𝒫 2 (𝐑) . (b) The differentiation operator ( the operator that takes 𝑝 to 𝑝 β€² ) on 𝒫 2 (𝐑) has an upper-triangular matrix with respect to the basis 1 , π‘₯ , π‘₯ 2 , which is not an orthonormal basis. Find the matrix of the differentiation operator on 𝒫 2 (𝐑) with respect to the orthonormal basis produced in (a) and verify that this matrix is upper triangular, as expected from the proof of 6.37. 9 Suppose 𝑒 1 , … , 𝑒 π‘š is the result of applying the Gram–Schmidt procedure to a linearly independent list 𝑣 1 , … , 𝑣 π‘š in 𝑉 . Prove that βŸ¨π‘£ π‘˜ , 𝑒 π‘˜ ⟩ > 0 for each π‘˜ = 1 , … , π‘š . 10 Suppose 𝑣 1 , … , 𝑣 π‘š is a linearly independent list in 𝑉 . Explain why the orthonormal list produced by the formulas of the Gram–Schmidt procedure (6.32) is the only orthonormal list 𝑒 1 , … , 𝑒 π‘š in 𝑉 such that βŸ¨π‘£ π‘˜ , 𝑒 π‘˜ ⟩ > 0 and span (𝑣 1 , … , 𝑣 π‘˜ ) = span (𝑒 1 , … , 𝑒 π‘˜ ) for each π‘˜ = 1 , … , π‘š . The result in this exercise is used in the proof of 7.58. 11 Find a polynomial π‘ž ∈ 𝒫 2 (𝐑) such that 𝑝( 12 ) = ∫ 10 π‘π‘ž for every 𝑝 ∈ 𝒫 2 (𝐑) . 12 Find a polynomial π‘ž ∈ 𝒫 2 (𝐑) such that ∫ 1 0 𝑝(π‘₯) cos (πœ‹π‘₯) 𝑑π‘₯ = ∫ 1 0 π‘π‘ž for every 𝑝 ∈ 𝒫 2 (𝐑) . Annotated Entity: ID: 222 Spans: True Boxes: True Text: Section 6B Orthonormal Bases 209 13 Show that a list 𝑣 1 , … , 𝑣 π‘š of vectors in 𝑉 is linearly dependent if and only if the Gram–Schmidt formula in 6.32 produces 𝑓 π‘˜ = 0 for some π‘˜ ∈ {1 , … , π‘š} . This exercise gives an alternative to Gaussian elimination techniques for determining whether a list of vectors in an inner product space is linearly dependent. 14 Suppose 𝑉 is a real inner product space and 𝑣 1 , … , 𝑣 π‘š is a linearly indepen- dent list of vectors in 𝑉 . Prove that there exist exactly 2 π‘š orthonormal lists 𝑒 1 , … , 𝑒 π‘š of vectors in 𝑉 such that span (𝑣 1 , … , 𝑣 π‘˜ ) = span (𝑒 1 , … , 𝑒 π‘˜ ) for all π‘˜ ∈ {1 , … , π‘š} . 15 Suppose βŸ¨β‹… , β‹…βŸ© 1 and βŸ¨β‹… , β‹…βŸ© 2 are inner products on 𝑉 such that βŸ¨π‘’ , π‘£βŸ© 1 = 0 if and only if βŸ¨π‘’ , π‘£βŸ© 2 = 0 . Prove that there is a positive number 𝑐 such that βŸ¨π‘’ , π‘£βŸ© 1 = π‘βŸ¨π‘’ , π‘£βŸ© 2 for every 𝑒 , 𝑣 ∈ 𝑉 . This exercise shows that if two inner products have the same pairs of orthogonal vectors, then each of the inner products is a scalar multiple of the other inner product. 16 Suppose 𝑉 is finite-dimensional. Suppose βŸ¨β‹… , β‹…βŸ© 1 , βŸ¨β‹… , β‹…βŸ© 2 are inner products on 𝑉 with corresponding norms β€–β‹…β€– 1 and β€–β‹…β€– 2 . Prove that there exists a positive number 𝑐 such that ‖𝑣‖ 1 ≀ 𝑐‖𝑣‖ 2 for every 𝑣 ∈ 𝑉 . 17 Suppose 𝐅 = 𝐂 and 𝑉 is finite-dimensional. Prove that if 𝑇 is an operator on 𝑉 such that 1 is the only eigenvalue of 𝑇 and ‖𝑇𝑣‖ ≀ ‖𝑣‖ for all 𝑣 ∈ 𝑉 , then 𝑇 is the identity operator. 18 Suppose 𝑒 1 , … , 𝑒 π‘š is a linearly independent list in 𝑉 . Show that there exists 𝑣 ∈ 𝑉 such that βŸ¨π‘’ π‘˜ , π‘£βŸ© = 1 for all π‘˜ ∈ {1 , … , π‘š} . 19 Suppose 𝑣 1 , … , 𝑣 𝑛 is a basis of 𝑉 . Prove that there exists a basis 𝑒 1 , … , 𝑒 𝑛 of 𝑉 such that βŸ¨π‘£ 𝑗 , 𝑒 π‘˜ ⟩ = ⎧{ ⎨{⎩ 0 if 𝑗 β‰  π‘˜ , 1 if 𝑗 = π‘˜. 20 Suppose 𝐅 = 𝐂 , 𝑉 is finite-dimensional, and β„° βŠ† β„’ (𝑉) is such that 𝑆𝑇 = 𝑇𝑆 for all 𝑆 , 𝑇 ∈ β„° . Prove that there is an orthonormal basis of 𝑉 with respect to which every element of β„° has an upper-triangular matrix. This exercise strengthens Exercise 9 ( b ) in Section 5E ( in the context of inner product spaces ) by asserting that the basis in that exercise can be chosen to be orthonormal. 21 Suppose 𝐅 = 𝐂 , 𝑉 is finite-dimensional, 𝑇 ∈ β„’ (𝑉) , and all eigenvalues of 𝑇 have absolute value less than 1 . Let πœ– > 0 . Prove that there exists a positive integer π‘š such that βˆ₯𝑇 π‘š 𝑣βˆ₯ ≀ πœ–β€–π‘£β€– for every 𝑣 ∈ 𝑉 . Annotated Entity: ID: 223 Spans: True Boxes: True Text: 210 Chapter 6 Inner Product Spaces 22 Suppose 𝐢[βˆ’1 , 1] is the vector space of continuous real-valued functions on the interval [βˆ’1 , 1] with inner product given by ⟨ 𝑓 , π‘”βŸ© = ∫ 1 βˆ’1 𝑓 𝑔 for all 𝑓 , 𝑔 ∈ 𝐢[βˆ’1 , 1] . Let πœ‘ be the linear functional on 𝐢[βˆ’1 , 1] defined by πœ‘( 𝑓 ) = 𝑓 (0) . Show that there does not exist 𝑔 ∈ 𝐢[βˆ’1 , 1] such that πœ‘( 𝑓 ) = ⟨ 𝑓 , π‘”βŸ© for every 𝑓 ∈ 𝐢[βˆ’1 , 1] . This exercise shows that the Riesz representation theorem ( 6.42 ) does not hold on infinite-dimensional vector spaces without additional hypotheses on 𝑉 and πœ‘ . 23 For all 𝑒 , 𝑣 ∈ 𝑉 , define 𝑑(𝑒 , 𝑣) = ‖𝑒 βˆ’ 𝑣‖ . (a) Show that 𝑑 is a metric on 𝑉 . (b) Show that if 𝑉 is finite-dimensional, then 𝑑 is a complete metric on 𝑉 (meaning that every Cauchy sequence converges). (c) Show that every finite-dimensional subspace of 𝑉 is a closed subset of 𝑉 (with respect to the metric 𝑑 ). This exercise requires familiarity with metric spaces. orthogonality at the Supreme Court Law professor Richard Friedman presenting a case before the U.S. Supreme Court in 2010: Mr. Friedman : I think that issue is entirely orthogonal to the issue here because the Commonwealth is acknowledgingβ€” Chief Justice Roberts : I’m sorry. Entirely what? Mr. Friedman : Orthogonal. Right angle. Unrelated. Irrelevant. Chief Justice Roberts : Oh. Justice Scalia : What was that adjective? I liked that. Mr. Friedman : Orthogonal. Chief Justice Roberts : Orthogonal. Mr. Friedman : Right, right. Justice Scalia : Orthogonal, ooh. (Laughter.) Justice Kennedy : I knew this case presented us a problem. (Laughter.) Annotated Entity: ID: 224 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 211 6C Orthogonal Complements and Minimization Problems Orthogonal Complements 6.46 definition: orthogonal complement, π‘ˆ βŸ‚ If π‘ˆ is a subset of 𝑉 , then the orthogonal complement of π‘ˆ , denoted by π‘ˆ βŸ‚ , is the set of all vectors in 𝑉 that are orthogonal to every vector in π‘ˆ : π‘ˆ βŸ‚ = {𝑣 ∈ 𝑉 ∢ βŸ¨π‘’ , π‘£βŸ© = 0 for every 𝑒 ∈ π‘ˆ}. The orthogonal complement π‘ˆ βŸ‚ depends on 𝑉 as well as on π‘ˆ . However, the inner product space 𝑉 should always be clear from the context and thus it can be omitted from the notation. 6.47 example: orthogonal complements β€’ If 𝑉 = 𝐑 3 and π‘ˆ is the subset of 𝑉 consisting of the single point (2 , 3 , 5) , then π‘ˆ βŸ‚ is the plane {(π‘₯ , 𝑦 , 𝑧) ∈ 𝐑 3 ∢ 2π‘₯ + 3𝑦 + 5𝑧 = 0} . β€’ If 𝑉 = 𝐑 3 and π‘ˆ is the plane {(π‘₯ , 𝑦 , 𝑧) ∈ 𝐑 3 ∢ 2π‘₯ + 3𝑦 + 5𝑧 = 0} , then π‘ˆ βŸ‚ is the line {(2𝑑 , 3𝑑 , 5𝑑) ∢ 𝑑 ∈ 𝐑} . β€’ More generally, if π‘ˆ is a plane in 𝐑 3 containing the origin, then π‘ˆ βŸ‚ is the line containing the origin that is perpendicular to π‘ˆ . β€’ If π‘ˆ is a line in 𝐑 3 containing the origin, then π‘ˆ βŸ‚ is the plane containing the origin that is perpendicular to π‘ˆ . β€’ If 𝑉 = 𝐅 5 and π‘ˆ = {(π‘Ž , 𝑏 , 0 , 0 , 0) ∈ 𝐅 5 ∢ π‘Ž , 𝑏 ∈ 𝐅} , then π‘ˆ βŸ‚ = {(0 , 0 , π‘₯ , 𝑦 , 𝑧) ∈ 𝐅 5 ∢ π‘₯ , 𝑦 , 𝑧 ∈ 𝐅}. β€’ If 𝑒 1 , … , 𝑒 π‘š , 𝑓 1 , … , 𝑓 𝑛 is an orthonormal basis of 𝑉 , then ( span (𝑒 1 , … , 𝑒 π‘š )) βŸ‚ = span ( 𝑓 1 , … , 𝑓 𝑛 ). We begin with some straightforward consequences of the definition. 6.48 properties of orthogonal complement (a) If π‘ˆ is a subset of 𝑉 , then π‘ˆ βŸ‚ is a subspace of 𝑉 . (b) {0} βŸ‚ = 𝑉 . (c) 𝑉 βŸ‚ = {0} . (d) If π‘ˆ is a subset of 𝑉 , then π‘ˆ ∩ π‘ˆ βŸ‚ βŠ† {0} . (e) If 𝐺 and 𝐻 are subsets of 𝑉 and 𝐺 βŠ† 𝐻 , then 𝐻 βŸ‚ βŠ† 𝐺 βŸ‚ . Annotated Entity: ID: 225 Spans: True Boxes: True Text: 212 Chapter 6 Inner Product Spaces Proof (a) Suppose π‘ˆ is a subset of 𝑉 . Then βŸ¨π‘’ , 0⟩ = 0 for every 𝑒 ∈ π‘ˆ ; thus 0 ∈ π‘ˆ βŸ‚ . Suppose 𝑣 , 𝑀 ∈ π‘ˆ βŸ‚ . If 𝑒 ∈ π‘ˆ , then βŸ¨π‘’ , 𝑣 + π‘€βŸ© = βŸ¨π‘’ , π‘£βŸ© + βŸ¨π‘’ , π‘€βŸ© = 0 + 0 = 0. Thus 𝑣 + 𝑀 ∈ π‘ˆ βŸ‚ , which shows that π‘ˆ βŸ‚ is closed under addition. Similarly, suppose πœ† ∈ 𝐅 and 𝑣 ∈ π‘ˆ βŸ‚ . If 𝑒 ∈ π‘ˆ , then βŸ¨π‘’ , πœ†π‘£βŸ© = πœ†βŸ¨π‘’ , π‘£βŸ© = πœ† β‹… 0 = 0. Thus πœ†π‘£ ∈ π‘ˆ βŸ‚ , which shows that π‘ˆ βŸ‚ is closed under scalar multiplication. Thus π‘ˆ βŸ‚ is a subspace of 𝑉 . (b) Suppose that 𝑣 ∈ 𝑉 . Then ⟨0 , π‘£βŸ© = 0 , which implies that 𝑣 ∈ {0} βŸ‚ . Thus {0} βŸ‚ = 𝑉 . (c) Suppose that 𝑣 ∈ 𝑉 βŸ‚ . Then βŸ¨π‘£ , π‘£βŸ© = 0 , which implies that 𝑣 = 0 . Thus 𝑉 βŸ‚ = {0} . (d) Suppose π‘ˆ is a subset of 𝑉 and 𝑒 ∈ π‘ˆ ∩ π‘ˆ βŸ‚ . Then βŸ¨π‘’ , π‘’βŸ© = 0 , which implies that 𝑒 = 0 . Thus π‘ˆ ∩ π‘ˆ βŸ‚ βŠ† {0} . (e) Suppose 𝐺 and 𝐻 are subsets of 𝑉 and 𝐺 βŠ† 𝐻 . Suppose 𝑣 ∈ 𝐻 βŸ‚ . Then βŸ¨π‘’ , π‘£βŸ© = 0 for every 𝑒 ∈ 𝐻 , which implies that βŸ¨π‘’ , π‘£βŸ© = 0 for every 𝑒 ∈ 𝐺 . Hence 𝑣 ∈ 𝐺 βŸ‚ . Thus 𝐻 βŸ‚ βŠ† 𝐺 βŸ‚ . Recall that if π‘ˆ and π‘Š are subspaces of 𝑉 , then 𝑉 is the direct sum of π‘ˆ and π‘Š (written 𝑉 = π‘ˆ βŠ• π‘Š ) if each element of 𝑉 can be written in exactly one way as a vector in π‘ˆ plus a vector in π‘Š (see 1.41). Furthermore, this happens if and only if 𝑉 = π‘ˆ + π‘Š and π‘ˆ ∩ π‘Š = {0} (see 1.46). The next result shows that every finite-dimensional subspace of 𝑉 leads to a natural direct sum decomposition of 𝑉 . See Exercise 16 for an example showing that the result below can fail without the hypothesis that the subspace π‘ˆ is finite- dimensional. 6.49 direct sum of a subspace and its orthogonal complement Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 . Then 𝑉 = π‘ˆ βŠ• π‘ˆ βŸ‚ . Proof First we will show that 𝑉 = π‘ˆ + π‘ˆ βŸ‚ . To do this, suppose that 𝑣 ∈ 𝑉 . Let 𝑒 1 , … , 𝑒 π‘š be an orthonormal basis of π‘ˆ . We want to write 𝑣 as the sum of a vector in π‘ˆ and a vector orthogonal to π‘ˆ . Annotated Entity: ID: 226 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 213 We have 6.50 𝑣 = βŸ¨π‘£ , 𝑒 1 βŸ©π‘’ 1 + β‹― + βŸ¨π‘£ , 𝑒 π‘š βŸ©π‘’ π‘š ⏟⏟⏟⏟⏟⏟⏟⏟⏟ 𝑒 + 𝑣 βˆ’ βŸ¨π‘£ , 𝑒 1 βŸ©π‘’ 1 βˆ’ β‹― βˆ’ βŸ¨π‘£ , 𝑒 π‘š βŸ©π‘’ π‘š ⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟⏟ 𝑀 . Let 𝑒 and 𝑀 be defined as in the equation above (as was done in the proof of 6.26). Because each 𝑒 π‘˜ ∈ π‘ˆ , we see that 𝑒 ∈ π‘ˆ . Because 𝑒 1 , … , 𝑒 π‘š is an orthonormal list, for each π‘˜ = 1 , … , π‘š we have βŸ¨π‘€ , 𝑒 π‘˜ ⟩ = βŸ¨π‘£ , 𝑒 π‘˜ ⟩ βˆ’ βŸ¨π‘£ , 𝑒 π‘˜ ⟩ = 0. Thus 𝑀 is orthogonal to every vector in span (𝑒 1 , … , 𝑒 π‘š ) , which shows that 𝑀 ∈ π‘ˆ βŸ‚ . Hence we have written 𝑣 = 𝑒 + 𝑀 , where 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ , completing the proof that 𝑉 = π‘ˆ + π‘ˆ βŸ‚ . From 6.48(d), we know that π‘ˆ ∩ π‘ˆ βŸ‚ = {0} . Now equation 𝑉 = π‘ˆ + π‘ˆ βŸ‚ implies that 𝑉 = π‘ˆ βŠ• π‘ˆ βŸ‚ (see 1.46). Now we can see how to compute dim π‘ˆ βŸ‚ from dim π‘ˆ . 6.51 dimension of orthogonal complement Suppose 𝑉 is finite-dimensional and π‘ˆ is a subspace of 𝑉 . Then dim π‘ˆ βŸ‚ = dim 𝑉 βˆ’ dim π‘ˆ. Proof The formula for dim π‘ˆ βŸ‚ follows immediately from 6.49 and 3.94. The next result is an important consequence of 6.49. 6.52 orthogonal complement of the orthogonal complement Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 . Then π‘ˆ = (π‘ˆ βŸ‚ ) βŸ‚ . Proof First we will show that 6.53 π‘ˆ βŠ† (π‘ˆ βŸ‚ ) βŸ‚ . To do this, suppose 𝑒 ∈ π‘ˆ . Then βŸ¨π‘’ , π‘€βŸ© = 0 for every 𝑀 ∈ π‘ˆ βŸ‚ ( by the definition of π‘ˆ βŸ‚ ) . Because 𝑒 is orthogonal to every vector in π‘ˆ βŸ‚ , we have 𝑒 ∈ (π‘ˆ βŸ‚ ) βŸ‚ , completing the proof of 6.53. To prove the inclusion in the other direction, suppose 𝑣 ∈ (π‘ˆ βŸ‚ ) βŸ‚ . By 6.49, we can write 𝑣 = 𝑒 + 𝑀 , where 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ . We have 𝑣 βˆ’ 𝑒 = 𝑀 ∈ π‘ˆ βŸ‚ . Because 𝑣 ∈ (π‘ˆ βŸ‚ ) βŸ‚ and 𝑒 ∈ (π‘ˆ βŸ‚ ) βŸ‚ (from 6.53), we have 𝑣 βˆ’ 𝑒 ∈ (π‘ˆ βŸ‚ ) βŸ‚ . Thus 𝑣 βˆ’ 𝑒 ∈ π‘ˆ βŸ‚ ∩ (π‘ˆ βŸ‚ ) βŸ‚ , which implies that 𝑣 βˆ’ 𝑒 = 0 [by 6.48(d)], which implies that 𝑣 = 𝑒 , which implies that 𝑣 ∈ π‘ˆ . Thus (π‘ˆ βŸ‚ ) βŸ‚ βŠ† π‘ˆ , which along with 6.53 completes the proof. Annotated Entity: ID: 227 Spans: True Boxes: True Text: 214 Chapter 6 Inner Product Spaces Exercise 16 ( a ) shows that the result below is not true without the hypothesis that π‘ˆ is finite-dimensional. Suppose π‘ˆ is a subspace of 𝑉 and we want to show that π‘ˆ = 𝑉 . In some situations, the easiest way to do this is to show that the only vector orthogonal to π‘ˆ is 0 , and then use the result below. For example, the result below is useful for Exercise 4. 6.54 π‘ˆ βŸ‚ = {0} ⟺ π‘ˆ = 𝑉 ( for π‘ˆ a finite-dimensional subspace of 𝑉 ) Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 . Then π‘ˆ βŸ‚ = {0} ⟺ π‘ˆ = 𝑉. Proof First suppose π‘ˆ βŸ‚ = {0} . Then by 6.52, π‘ˆ = (π‘ˆ βŸ‚ ) βŸ‚ = {0} βŸ‚ = 𝑉 , as desired. Conversely, if π‘ˆ = 𝑉 , then π‘ˆ βŸ‚ = 𝑉 βŸ‚ = {0} by 6.48(c). We now define an operator 𝑃 π‘ˆ for each finite-dimensional subspace π‘ˆ of 𝑉 . 6.55 definition: orthogonal projection, 𝑃 π‘ˆ Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 . The orthogonal projection of 𝑉 onto π‘ˆ is the operator 𝑃 π‘ˆ ∈ β„’ (𝑉) defined as follows: For each 𝑣 ∈ 𝑉 , write 𝑣 = 𝑒 + 𝑀 , where 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ . Then let 𝑃 π‘ˆ 𝑣 = 𝑒 . The direct sum decomposition 𝑉 = π‘ˆ βŠ• π‘ˆ βŸ‚ given by 6.49 shows that each 𝑣 ∈ 𝑉 can be uniquely written in the form 𝑣 = 𝑒 + 𝑀 with 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ . Thus 𝑃 π‘ˆ 𝑣 is well defined. See the figure that accompanies the proof of 6.61 for the picture describing 𝑃 π‘ˆ 𝑣 that you should keep in mind. 6.56 example: orthogonal projection onto one-dimensional subspace Suppose 𝑒 ∈ 𝑉 with 𝑒 β‰  0 and π‘ˆ is the one-dimensional subspace of 𝑉 defined by π‘ˆ = span (𝑒) . If 𝑣 ∈ 𝑉 , then 𝑣 = βŸ¨π‘£ , π‘’βŸ© ‖𝑒‖ 2 𝑒 + (𝑣 βˆ’ βŸ¨π‘£ , π‘’βŸ© ‖𝑒‖ 2 𝑒) , where the first term on the right is in span (𝑒) (and thus is in π‘ˆ ) and the second term on the right is orthogonal to 𝑒 (and thus is in π‘ˆ βŸ‚ ) . Thus 𝑃 π‘ˆ 𝑣 equals the first term on the right. In other words, we have the formula 𝑃 π‘ˆ 𝑣 = βŸ¨π‘£ , π‘’βŸ© ‖𝑒‖ 2 𝑒 for every 𝑣 ∈ 𝑉 . Taking 𝑣 = 𝑒 , the formula above becomes 𝑃 π‘ˆ 𝑒 = 𝑒 , as expected. Furthermore, taking 𝑣 ∈ {𝑒} βŸ‚ , the formula above becomes 𝑃 π‘ˆ 𝑣 = 0 , also as expected. Annotated Entity: ID: 228 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 215 6.57 properties of orthogonal projection 𝑃 π‘ˆ Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 . Then (a) 𝑃 π‘ˆ ∈ β„’ (𝑉) ; (b) 𝑃 π‘ˆ 𝑒 = 𝑒 for every 𝑒 ∈ π‘ˆ ; (c) 𝑃 π‘ˆ 𝑀 = 0 for every 𝑀 ∈ π‘ˆ βŸ‚ ; (d) range 𝑃 π‘ˆ = π‘ˆ ; (e) null 𝑃 π‘ˆ = π‘ˆ βŸ‚ ; (f) 𝑣 βˆ’ 𝑃 π‘ˆ 𝑣 ∈ π‘ˆ βŸ‚ for every 𝑣 ∈ 𝑉 ; (g) 𝑃 π‘ˆ2 = 𝑃 π‘ˆ ; (h) ‖𝑃 π‘ˆ 𝑣‖ ≀ ‖𝑣‖ for every 𝑣 ∈ 𝑉 ; (i) if 𝑒 1 , … , 𝑒 π‘š is an orthonormal basis of π‘ˆ and 𝑣 ∈ 𝑉 , then 𝑃 π‘ˆ 𝑣 = βŸ¨π‘£ , 𝑒 1 βŸ©π‘’ 1 + β‹― + βŸ¨π‘£ , 𝑒 π‘š βŸ©π‘’ π‘š . Proof (a) To show that 𝑃 π‘ˆ is a linear map on 𝑉 , suppose 𝑣 1 , 𝑣 2 ∈ 𝑉 . Write 𝑣 1 = 𝑒 1 + 𝑀 1 and 𝑣 2 = 𝑒 2 + 𝑀 2 with 𝑒 1 , 𝑒 2 ∈ π‘ˆ and 𝑀 1 , 𝑀 2 ∈ π‘ˆ βŸ‚ . Thus 𝑃 π‘ˆ 𝑣 1 = 𝑒 1 and 𝑃 π‘ˆ 𝑣 2 = 𝑒 2 . Now 𝑣 1 + 𝑣 2 = (𝑒 1 + 𝑒 2 ) + (𝑀 1 + 𝑀 2 ) , where 𝑒 1 + 𝑒 2 ∈ π‘ˆ and 𝑀 1 + 𝑀 2 ∈ π‘ˆ βŸ‚ . Thus 𝑃 π‘ˆ (𝑣 1 + 𝑣 2 ) = 𝑒 1 + 𝑒 2 = 𝑃 π‘ˆ 𝑣 1 + 𝑃 π‘ˆ 𝑣 2 . Similarly, suppose πœ† ∈ 𝐅 and 𝑣 ∈ 𝑉 . Write 𝑣 = 𝑒 + 𝑀 , where 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ . Then πœ†π‘£ = πœ†π‘’ + πœ†π‘€ with πœ†π‘’ ∈ π‘ˆ and πœ†π‘€ ∈ π‘ˆ βŸ‚ . Thus 𝑃 π‘ˆ (πœ†π‘£) = πœ†π‘’ = πœ†π‘ƒ π‘ˆ 𝑣 . Hence 𝑃 π‘ˆ is a linear map from 𝑉 to 𝑉 . (b) Suppose 𝑒 ∈ π‘ˆ . We can write 𝑒 = 𝑒 + 0 , where 𝑒 ∈ π‘ˆ and 0 ∈ π‘ˆ βŸ‚ . Thus 𝑃 π‘ˆ 𝑒 = 𝑒 . (c) Suppose 𝑀 ∈ π‘ˆ βŸ‚ . We can write 𝑀 = 0 + 𝑀 , where 0 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ . Thus 𝑃 π‘ˆ 𝑀 = 0 . (d) The definition of 𝑃 π‘ˆ implies that range 𝑃 π‘ˆ βŠ† π‘ˆ . Furthermore, (b) implies that π‘ˆ βŠ† range 𝑃 π‘ˆ . Thus range 𝑃 π‘ˆ = π‘ˆ . (e) The inclusion π‘ˆ βŸ‚ βŠ† null 𝑃 π‘ˆ follows from (c). To prove the inclusion in the other direction, note that if 𝑣 ∈ null 𝑃 π‘ˆ then the decomposition given by 6.49 must be 𝑣 = 0 + 𝑣 , where 0 ∈ π‘ˆ and 𝑣 ∈ π‘ˆ βŸ‚ . Thus null 𝑃 π‘ˆ βŠ† π‘ˆ βŸ‚ . Annotated Entity: ID: 229 Spans: True Boxes: True Text: 216 Chapter 6 Inner Product Spaces (f) If 𝑣 ∈ 𝑉 and 𝑣 = 𝑒 + 𝑀 with 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ , then 𝑣 βˆ’ 𝑃 π‘ˆ 𝑣 = 𝑣 βˆ’ 𝑒 = 𝑀 ∈ π‘ˆ βŸ‚ . (g) If 𝑣 ∈ 𝑉 and 𝑣 = 𝑒 + 𝑀 with 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ , then (𝑃 π‘ˆ2 )𝑣 = 𝑃 π‘ˆ (𝑃 π‘ˆ 𝑣) = 𝑃 π‘ˆ 𝑒 = 𝑒 = 𝑃 π‘ˆ 𝑣. (h) If 𝑣 ∈ 𝑉 and 𝑣 = 𝑒 + 𝑀 with 𝑒 ∈ π‘ˆ and 𝑀 ∈ π‘ˆ βŸ‚ , then ‖𝑃 π‘ˆ 𝑣‖ 2 = ‖𝑒‖ 2 ≀ ‖𝑒‖ 2 + ‖𝑀‖ 2 = ‖𝑣‖ 2 , where the last equality comes from the Pythagorean theorem. (i) The formula for 𝑃 π‘ˆ 𝑣 follows from equation 6.50 in the proof of 6.49. In the previous section we proved the Riesz representation theorem (6.42), whose key part states that every linear functional on a finite-dimensional inner product space is given by taking the inner product with some fixed vector. Seeing a different proof often provides new insight. Thus we now give a new proof of the key part of the Riesz representation theorem using orthogonal complements instead of orthonormal bases as in our previous proof. The restatement below of the Riesz representation theorem provides an iden- tification of 𝑉 with 𝑉 β€² . We will prove only the β€œonto” part of the result below because the more routine β€œone-to-one” part of the result can be proved as in 6.42. Intuition behind this new proof: If πœ‘ ∈ 𝑉 β€² , 𝑣 ∈ 𝑉 , and πœ‘(𝑒) = βŸ¨π‘’ , π‘£βŸ© for all 𝑒 ∈ 𝑉 , then 𝑣 ∈ ( null πœ‘) βŸ‚ . However, ( null πœ‘) βŸ‚ is a one-dimensional subspace of 𝑉 (except for the trivial case in which πœ‘ = 0 ), as follows from 6.51 and 3.21. Thus we can obtain 𝑣 be choosing any nonzero element of ( null πœ‘) βŸ‚ and then multiplying by an appropriate scalar, as is done in the proof below. 6.58 Riesz representation theorem, revisited Suppose 𝑉 is finite-dimensional. For each 𝑣 ∈ 𝑉 , define πœ‘ 𝑣 ∈ 𝑉 β€² by πœ‘ 𝑣 (𝑒) = βŸ¨π‘’ , π‘£βŸ© for each 𝑒 ∈ 𝑉 . Then 𝑣 ↦ πœ‘ 𝑣 is a one-to-one function from 𝑉 onto 𝑉 β€² . Caution: The function 𝑣 ↦ πœ‘ 𝑣 is a linear mapping from 𝑉 to 𝑉 β€² if 𝐅 = 𝐑 . However, this function is not linear if 𝐅 = 𝐂 because πœ‘ πœ†π‘£ = πœ†πœ‘ 𝑣 if πœ† ∈ 𝐂 . Proof To show that 𝑣 ↦ πœ‘ 𝑣 is surjective, suppose πœ‘ ∈ 𝑉 β€² . If πœ‘ = 0 , then πœ‘ = πœ‘ 0 . Thus assume πœ‘ β‰  0 . Hence null πœ‘ β‰  𝑉 , which implies that ( null πœ‘) βŸ‚ β‰  {0} (by 6.49 with π‘ˆ = null πœ‘ ). Let 𝑀 ∈ ( null πœ‘) βŸ‚ be such that 𝑀 β‰  0 . Let 6.59 𝑣 = πœ‘(𝑀) ‖𝑀‖ 2 𝑀. Then 𝑣 ∈ ( null πœ‘) βŸ‚ . Also, 𝑣 β‰  0 (because 𝑀 βˆ‰ null πœ‘ ). Annotated Entity: ID: 230 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 217 Taking the norm of both sides of 6.59 gives 6.60 ‖𝑣‖ = |πœ‘(𝑀)| ‖𝑀‖ . Applying πœ‘ to both sides of 6.59 and then using 6.60, we have πœ‘(𝑣) = |πœ‘(𝑀)| 2 ‖𝑀‖ 2 = ‖𝑣‖ 2 . Now suppose 𝑒 ∈ 𝑉 . Using the equation above, we have 𝑒 = (𝑒 βˆ’ πœ‘(𝑒) πœ‘(𝑣)𝑣) + πœ‘(𝑒) ‖𝑣‖ 2 𝑣. The first term in parentheses above is in null πœ‘ and hence is orthogonal to 𝑣 . Thus taking the inner product of both sides of the equation above with 𝑣 shows that βŸ¨π‘’ , π‘£βŸ© = πœ‘(𝑒) ‖𝑣‖ 2 βŸ¨π‘£ , π‘£βŸ© = πœ‘(𝑒). Thus πœ‘ = πœ‘ 𝑣 , showing that 𝑣 ↦ πœ‘ 𝑣 is surjective, as desired. See Exercise 13 for yet another proof of the Riesz representation theorem. Minimization Problems The remarkable simplicity of the solu- tion to this minimization problem has led to many important applications of inner product spaces outside of pure mathematics. The following problem often arises: Given a subspace π‘ˆ of 𝑉 and a point 𝑣 ∈ 𝑉 , find a point 𝑒 ∈ π‘ˆ such that ‖𝑣 βˆ’ 𝑒‖ is as small as possible. The next result shows that 𝑒 = 𝑃 π‘ˆ 𝑣 is the unique solution of this minimization problem. 6.61 minimizing distance to a subspace Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 , 𝑣 ∈ 𝑉 , and 𝑒 ∈ π‘ˆ . Then ‖𝑣 βˆ’ 𝑃 π‘ˆ 𝑣‖ ≀ ‖𝑣 βˆ’ 𝑒‖. Furthermore, the inequality above is an equality if and only if 𝑒 = 𝑃 π‘ˆ 𝑣 . Proof We have ‖𝑣 βˆ’ 𝑃 π‘ˆ 𝑣‖ 2 ≀ ‖𝑣 βˆ’ 𝑃 π‘ˆ 𝑣‖ 2 + ‖𝑃 π‘ˆ 𝑣 βˆ’ 𝑒‖ 2 6.62 = βˆ₯(𝑣 βˆ’ 𝑃 π‘ˆ 𝑣) + (𝑃 π‘ˆ 𝑣 βˆ’ 𝑒)βˆ₯ 2 = ‖𝑣 βˆ’ 𝑒‖ 2 , Annotated Entity: ID: 231 Spans: True Boxes: True Text: 218 Chapter 6 Inner Product Spaces 𝑃 π‘ˆ 𝑣 is the closest point in π‘ˆ to 𝑣 . where the first line above holds because 0 ≀ ‖𝑃 π‘ˆ 𝑣 βˆ’ 𝑒‖ 2 , the second line above comes from the Pythagorean the- orem [ which applies because 𝑣 βˆ’ 𝑃 π‘ˆ 𝑣 ∈ π‘ˆ βŸ‚ by 6.57(f), and 𝑃 π‘ˆ 𝑣 βˆ’ 𝑒 ∈ π‘ˆ] , and the third line above holds by simple algebra. Taking square roots gives the desired inequality. The inequality proved above is an equality if and only if 6.62 is an equality, which happens if and only if ‖𝑃 π‘ˆ 𝑣 βˆ’ 𝑒‖ = 0 , which happens if and only if 𝑒 = 𝑃 π‘ˆ 𝑣 . The last result is often combined with the formula 6.57(i) to compute explicit solutions to minimization problems, as in the following example. 6.63 example: using linear algebra to approximate the sine function Suppose we want to find a polynomial 𝑒 with real coefficients and of degree at most 5 that approximates the sine function as well as possible on the interval [βˆ’πœ‹ , πœ‹] , in the sense that ∫ πœ‹ βˆ’πœ‹ ∣ sin π‘₯ βˆ’ 𝑒(π‘₯)∣ 2 𝑑π‘₯ is as small as possible. Let 𝐢[βˆ’πœ‹ , πœ‹] denote the real inner product space of continuous real-valued functions on [βˆ’πœ‹ , πœ‹] with inner product 6.64 ⟨ 𝑓 , π‘”βŸ© = ∫ πœ‹ βˆ’πœ‹ 𝑓 𝑔. Let 𝑣 ∈ 𝐢[βˆ’πœ‹ , πœ‹] be the function defined by 𝑣(π‘₯) = sin π‘₯ . Let π‘ˆ denote the subspace of 𝐢[βˆ’πœ‹ , πœ‹] consisting of the polynomials with real coefficients and of degree at most 5 . Our problem can now be reformulated as follows: Find 𝑒 ∈ π‘ˆ such that ‖𝑣 βˆ’ 𝑒‖ is as small as possible. A computer that can integrate is useful here. To compute the solution to our ap- proximation problem, first apply the Gram–Schmidt procedure (using the in- ner product given by 6.64) to the basis 1 , π‘₯ , π‘₯ 2 , π‘₯ 3 , π‘₯ 4 , π‘₯ 5 of π‘ˆ , producing an ortho- normal basis 𝑒 1 , 𝑒 2 , 𝑒 3 , 𝑒 4 , 𝑒 5 , 𝑒 6 of π‘ˆ . Then, again using the inner product given by 6.64, compute 𝑃 π‘ˆ 𝑣 using 6.57(i) (with π‘š = 6 ). Doing this computation shows that 𝑃 π‘ˆ 𝑣 is the function 𝑒 defined by 6.65 𝑒(π‘₯) = 0.987862π‘₯ βˆ’ 0.155271π‘₯ 3 + 0.00564312π‘₯ 5 , where the πœ‹ ’s that appear in the exact answer have been replaced with a good decimal approximation. By 6.61, the polynomial 𝑒 above is the best approximation to the sine function on [βˆ’πœ‹ , πœ‹] using polynomials of degree at most 5 ( here β€œbest approximation” means in the sense of minimizing ∫ πœ‹ βˆ’πœ‹ | sin π‘₯ βˆ’ 𝑒(π‘₯)| 2 𝑑π‘₯) . Annotated Entity: ID: 232 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 219 To see how good this approximation is, the next figure shows the graphs of both the sine function and our approximation 𝑒 given by 6.65 over the interval [βˆ’πœ‹ , πœ‹] . Graphs on [βˆ’πœ‹ , πœ‹] of the sine function ( red ) and its best fifth degree polynomial approximation 𝑒 ( blue ) from 6.65. Our approximation 6.65 is so accurate that the two graphs are almost identicalβ€” our eyes may see only one graph! Here the red graph is placed almost exactly over the blue graph. If you are viewing this on an electronic device, enlarge the picture above by 400% near πœ‹ or βˆ’πœ‹ to see a small gap between the two graphs. Another well-known approximation to the sine function by a polynomial of degree 5 is given by the Taylor polynomial 𝑝 defined by 6.66 𝑝(π‘₯) = π‘₯ βˆ’ π‘₯ 3 3! + π‘₯ 5 5! . To see how good this approximation is, the next picture shows the graphs of both the sine function and the Taylor polynomial 𝑝 over the interval [βˆ’πœ‹ , πœ‹] . Graphs on [βˆ’πœ‹ , πœ‹] of the sine function ( red ) and the Taylor polynomial ( blue ) from 6.66. The Taylor polynomial is an excellent approximation to sin π‘₯ for π‘₯ near 0 . But the picture above shows that for |π‘₯| > 2 , the Taylor polynomial is not so accurate, especially compared to 6.65. For example, taking π‘₯ = 3 , our approximation 6.65 estimates sin 3 with an error of approximately 0.001 , but the Taylor series 6.66 estimates sin 3 with an error of approximately 0.4 . Thus at π‘₯ = 3 , the error in the Taylor series is hundreds of times larger than the error given by 6.65. Linear algebra has helped us discover an approximation to the sine function that improves upon what we learned in calculus! Annotated Entity: ID: 233 Spans: True Boxes: True Text: 220 Chapter 6 Inner Product Spaces Pseudoinverse Suppose 𝑇 ∈ β„’ (𝑉 , π‘Š) and 𝑀 ∈ π‘Š . Consider the problem of finding 𝑣 ∈ 𝑉 such that 𝑇𝑣 = 𝑀. For example, if 𝑉 = 𝐅 𝑛 and π‘Š = 𝐅 π‘š , then the equation above could represent a system of π‘š linear equations in 𝑛 unknowns 𝑣 1 , … , 𝑣 𝑛 , where 𝑣 = (𝑣 1 , … , 𝑣 𝑛 ) . If 𝑇 is invertible, then the unique solution to the equation above is 𝑣 = 𝑇 βˆ’1 𝑀 . However, if 𝑇 is not invertible, then for some 𝑀 ∈ π‘Š there may not exist any solutions of the equation above, and for some 𝑀 ∈ π‘Š there may exist infinitely many solutions of the equation above. If 𝑇 is not invertible, then we can still try to do as well as possible with the equation above. For example, if the equation above has no solutions, then instead of solving the equation 𝑇𝑣 βˆ’ 𝑀 = 0 , we can try to find 𝑣 ∈ 𝑉 such that ‖𝑇𝑣 βˆ’ 𝑀‖ is as small as possible. As another example, if the equation above has infinitely many solutions 𝑣 ∈ 𝑉 , then among all those solutions we can try to find one such that ‖𝑣‖ is as small as possible. The pseudoinverse will provide the tool to solve the equation above as well as possible, even when 𝑇 is not invertible. We need the next result to define the pseudoinverse. In the next two proofs, we will use without further comment the result that if 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) , then null 𝑇 , ( null 𝑇) βŸ‚ , and range 𝑇 are all finite-dimensional. 6.67 restriction of a linear map to obtain a one-to-one and onto map Suppose 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . Then 𝑇| ( null 𝑇) βŸ‚ is an injective map of ( null 𝑇) βŸ‚ onto range 𝑇 . Proof Suppose that 𝑣 ∈ ( null 𝑇) βŸ‚ and 𝑇| ( null 𝑇) βŸ‚ 𝑣 = 0 . Hence 𝑇𝑣 = 0 and thus 𝑣 ∈ ( null 𝑇) ∩ ( null 𝑇) βŸ‚ , which implies that 𝑣 = 0 [ by 6.48(d) ] . Hence null 𝑇| ( null 𝑇) βŸ‚ = {0} , which implies that 𝑇| ( null 𝑇) βŸ‚ is injective, as desired. Clearly range 𝑇| ( null 𝑇) βŸ‚ βŠ† range 𝑇 . To prove the inclusion in the other direction, suppose 𝑀 ∈ range 𝑇 . Hence there exists 𝑣 ∈ 𝑉 such that 𝑀 = 𝑇𝑣 . There exist 𝑒 ∈ null 𝑇 and π‘₯ ∈ ( null 𝑇) βŸ‚ such that 𝑣 = 𝑒 + π‘₯ (by 6.49). Now 𝑇| ( null 𝑇) βŸ‚ π‘₯ = 𝑇π‘₯ = 𝑇𝑣 βˆ’ 𝑇𝑒 = 𝑀 βˆ’ 0 = 𝑀 , which shows that 𝑀 ∈ range 𝑇| ( null 𝑇) βŸ‚ . Hence range 𝑇 βŠ† range 𝑇| ( null 𝑇) βŸ‚ , complet- ing the proof that range 𝑇| ( null 𝑇) βŸ‚ = range 𝑇 . To produce the pseudoinverse notation 𝑇 † in TEX, type T Μ‚\dagger . Now we can define the pseudoinverse 𝑇 † (pronounced β€œ 𝑇 dagger”) of a linear map 𝑇 . In the next definition (and from now on), think of 𝑇| ( null 𝑇) βŸ‚ as an invertible linear map from ( null 𝑇) βŸ‚ onto range 𝑇 , as is justified by the result above. Annotated Entity: ID: 234 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 221 6.68 definition: pseudoinverse, 𝑇 † Suppose that 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . The pseudoinverse 𝑇 † ∈ β„’ (π‘Š , 𝑉) of 𝑇 is the linear map from π‘Š to 𝑉 defined by 𝑇 † 𝑀 = (𝑇| ( null 𝑇) βŸ‚ ) βˆ’1 𝑃 range 𝑇 𝑀 for each 𝑀 ∈ π‘Š . Recall that 𝑃 range 𝑇 𝑀 = 0 if 𝑀 ∈ ( range 𝑇) βŸ‚ and 𝑃 range 𝑇 𝑀 = 𝑀 if 𝑀 ∈ range 𝑇 . Thus if 𝑀 ∈ ( range 𝑇) βŸ‚ , then 𝑇 † 𝑀 = 0 , and if 𝑀 ∈ range 𝑇 , then 𝑇 † 𝑀 is the unique element of ( null 𝑇) βŸ‚ such that 𝑇(𝑇 † 𝑀) = 𝑀 . The pseudoinverse behaves much like an inverse, as we will see. 6.69 algebraic properties of the pseudoinverse Suppose 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . (a) If 𝑇 is invertible, then 𝑇 † = 𝑇 βˆ’1 . (b) 𝑇𝑇 † = 𝑃 range 𝑇 = the orthogonal projection of π‘Š onto range 𝑇 . (c) 𝑇 † 𝑇 = 𝑃 ( null 𝑇) βŸ‚ = the orthogonal projection of 𝑉 onto ( null 𝑇) βŸ‚ . Proof (a) Suppose 𝑇 is invertible. Then ( null 𝑇) βŸ‚ = 𝑉 and range 𝑇 = π‘Š . Thus 𝑇| ( null 𝑇) βŸ‚ = 𝑇 and 𝑃 range 𝑇 is the identity operator on π‘Š . Hence 𝑇 † = 𝑇 βˆ’1 . (b) Suppose 𝑀 ∈ range 𝑇 . Thus 𝑇𝑇 † 𝑀 = 𝑇(𝑇| ( null 𝑇) βŸ‚ ) βˆ’1 𝑀 = 𝑀 = 𝑃 range 𝑇 𝑀. If 𝑀 ∈ ( range 𝑇) βŸ‚ , then 𝑇 † 𝑀 = 0 . Hence 𝑇𝑇 † 𝑀 = 0 = 𝑃 range 𝑇 𝑀 . Thus 𝑇𝑇 † and 𝑃 range 𝑇 agree on range 𝑇 and on ( range 𝑇) βŸ‚ . Hence these two linear maps are equal (by 6.49). (c) Suppose 𝑣 ∈ ( null 𝑇) βŸ‚ . Because 𝑇𝑣 ∈ range 𝑇 , the definition of 𝑇 † shows that 𝑇 † (𝑇𝑣) = (𝑇| ( null 𝑇) βŸ‚ ) βˆ’1 (𝑇𝑣) = 𝑣 = 𝑃 ( null 𝑇) βŸ‚ 𝑣. If 𝑣 ∈ null 𝑇 , then 𝑇 † 𝑇𝑣 = 0 = 𝑃 ( null 𝑇) βŸ‚ 𝑣 . Thus 𝑇 † 𝑇 and 𝑃 ( null 𝑇) βŸ‚ agree on ( null 𝑇) βŸ‚ and on null 𝑇 . Hence these two linear maps are equal (by 6.49). The pseudoinverse is also called the Moore–Penrose inverse. Suppose that 𝑇 ∈ β„’ (𝑉 , π‘Š) . If 𝑇 is surjective, then 𝑇𝑇 † is the identity opera- tor on π‘Š , as follows from (b) in the result above. If 𝑇 is injective, then 𝑇 † 𝑇 is the identity operator on 𝑉 , as follows from (c) in the result above. For additional algebraic properties of the pseudoinverse, see Exercises 19–23. Annotated Entity: ID: 235 Spans: True Boxes: True Text: 222 Chapter 6 Inner Product Spaces For 𝑇 ∈ β„’ (𝑉 , π‘Š) and 𝑀 ∈ π‘Š , we now return to the problem of finding 𝑣 ∈ 𝑉 that solves the equation 𝑇𝑣 = 𝑀. As we noted earlier, if 𝑇 is invertible, then 𝑣 = 𝑇 βˆ’1 𝑀 is the unique solution, but if 𝑇 is not invertible, then 𝑇 βˆ’1 is not defined. However, the pseudoinverse 𝑇 † is defined. Taking 𝑣 = 𝑇 † 𝑀 makes 𝑇𝑣 as close to 𝑀 as possible, as shown by (a) of the next result. Thus the pseudoinverse provides what is called a best fit to the equation above. Among all vectors 𝑣 ∈ 𝑉 that make 𝑇𝑣 as close as possible to 𝑀 , the vector 𝑇 † 𝑀 has the smallest norm, as shown by combining (b) in the next result with the condition for equality in (a). 6.70 pseudoinverse provides best approximate solution or best solution Suppose 𝑉 is finite-dimensional, 𝑇 ∈ β„’ (𝑉 , π‘Š) , and 𝑀 ∈ π‘Š . (a) If 𝑣 ∈ 𝑉 , then βˆ₯𝑇(𝑇 † 𝑀) βˆ’ 𝑀βˆ₯ ≀ ‖𝑇𝑣 βˆ’ 𝑀‖ , with equality if and only if 𝑣 ∈ 𝑇 † 𝑀 + null 𝑇 . (b) If 𝑣 ∈ 𝑇 † 𝑀 + null 𝑇 , then βˆ₯𝑇 † 𝑀βˆ₯ ≀ ‖𝑣‖ , with equality if and only if 𝑣 = 𝑇 † 𝑀 . Proof (a) Suppose 𝑣 ∈ 𝑉 . Then 𝑇𝑣 βˆ’ 𝑀 = (𝑇𝑣 βˆ’ 𝑇𝑇 † 𝑀) + (𝑇𝑇 † 𝑀 βˆ’ 𝑀). The first term in parentheses above is in range 𝑇 . Because the operator 𝑇𝑇 † is the orthogonal projection of π‘Š onto range 𝑇 [ by 6.69(b) ] , the second term in parentheses above is in ( range 𝑇) βŸ‚ [ see 6.57(f) ] . Thus the Pythagorean theorem implies the desired inequality that the norm of the second term in parentheses above is less than or equal to ‖𝑇𝑣 βˆ’ 𝑀‖ , with equality if and only if the first term in parentheses above equals 0 . Hence we have equality if and only if 𝑣 βˆ’ 𝑇 † 𝑀 ∈ null 𝑇 , which is equivalent to the statement that 𝑣 ∈ 𝑇 † 𝑀 + null 𝑇 , completing the proof of (a). (b) Suppose 𝑣 ∈ 𝑇 † 𝑀 + null 𝑇 . Hence 𝑣 βˆ’ 𝑇 † 𝑀 ∈ null 𝑇 . Now 𝑣 = (𝑣 βˆ’ 𝑇 † 𝑀) + 𝑇 † 𝑀. The definition of 𝑇 † implies that 𝑇 † 𝑀 ∈ ( null 𝑇) βŸ‚ . Thus the Pythagorean theorem implies that βˆ₯𝑇 † 𝑀βˆ₯ ≀ ‖𝑣‖ , with equality if and only if 𝑣 = 𝑇 † 𝑀 . A formula for 𝑇 † will be given in the next chapter (see 7.78). Annotated Entity: ID: 236 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 223 6.71 example: pseudoinverse of a linear map from 𝐅 4 to 𝐅 3 Suppose 𝑇 ∈ β„’ (𝐅 4 , 𝐅 3 ) is defined by 𝑇(π‘Ž , 𝑏 , 𝑐 , 𝑑) = (π‘Ž + 𝑏 + 𝑐 , 2𝑐 + 𝑑 , 0). This linear map is neither injective nor surjective, but we can compute its pseudo- inverse. To do this, first note that range 𝑇 = {(π‘₯ , 𝑦 , 0) ∢ π‘₯ , 𝑦 ∈ 𝐅} . Thus 𝑃 range 𝑇 (π‘₯ , 𝑦 , 𝑧) = (π‘₯ , 𝑦 , 0) for each (π‘₯ , 𝑦 , 𝑧) ∈ 𝐅 3 . Also, null 𝑇 = {(π‘Ž , 𝑏 , 𝑐 , 𝑑) ∈ 𝐅 4 ∢ π‘Ž + 𝑏 + 𝑐 = 0 and 2𝑐 + 𝑑 = 0}. The list (βˆ’1 , 1 , 0 , 0) , (βˆ’1 , 0 , 1 , βˆ’2) of two vectors in null 𝑇 spans null 𝑇 because if (π‘Ž , 𝑏 , 𝑐 , 𝑑) ∈ null 𝑇 then (π‘Ž , 𝑏 , 𝑐 , 𝑑) = 𝑏(βˆ’1 , 1 , 0 , 0) + 𝑐(βˆ’1 , 0 , 1 , βˆ’2). Because the list (βˆ’1 , 1 , 0 , 0) , (βˆ’1 , 0 , 1 , βˆ’2) is linearly independent, this list is a basis of null 𝑇 . Now suppose (π‘₯ , 𝑦 , 𝑧) ∈ 𝐅 3 . Then 6.72 𝑇 † (π‘₯ , 𝑦 , 𝑧) = (𝑇| ( null 𝑇) βŸ‚ ) βˆ’1 𝑃 range 𝑇 (π‘₯ , 𝑦 , 𝑧) = (𝑇| ( null 𝑇) βŸ‚ ) βˆ’1 (π‘₯ , 𝑦 , 0). The right side of the equation above is the vector (π‘Ž , 𝑏 , 𝑐 , 𝑑) ∈ 𝐅 4 such that 𝑇(π‘Ž , 𝑏 , 𝑐 , 𝑑) = (π‘₯ , 𝑦 , 0) and (π‘Ž , 𝑏 , 𝑐 , 𝑑) ∈ ( null 𝑇) βŸ‚ . In other words, π‘Ž , 𝑏 , 𝑐 , 𝑑 must satisfy the following equations: π‘Ž + 𝑏 + 𝑐 = π‘₯ 2𝑐 + 𝑑 = 𝑦 βˆ’π‘Ž + 𝑏 = 0 βˆ’π‘Ž + 𝑐 βˆ’ 2𝑑 = 0 , where the first two equations are equivalent to the equation 𝑇(π‘Ž , 𝑏 , 𝑐 , 𝑑) = (π‘₯ , 𝑦 , 0) and the last two equations come from the condition for (π‘Ž , 𝑏 , 𝑐 , 𝑑) to be orthogo- nal to each of the basis vectors (βˆ’1 , 1 , 0 , 0) , (βˆ’1 , 0 , 1 , βˆ’2) in this basis of null 𝑇 . Thinking of π‘₯ and 𝑦 as constants and π‘Ž , 𝑏 , 𝑐 , 𝑑 as unknowns, we can solve the system above of four equations in four unknowns, getting π‘Ž = 111 (5π‘₯ βˆ’ 2𝑦) , 𝑏 = 111 (5π‘₯ βˆ’ 2𝑦) , 𝑐 = 111 (π‘₯ + 4𝑦) , 𝑑 = 111 (βˆ’2π‘₯ + 3𝑦). Hence 6.72 tells us that 𝑇 † (π‘₯ , 𝑦 , 𝑧) = 111 (5π‘₯ βˆ’ 2𝑦 , 5π‘₯ βˆ’ 2𝑦 , π‘₯ + 4𝑦 , βˆ’2π‘₯ + 3𝑦). The formula above for 𝑇 † shows that 𝑇𝑇 † (π‘₯ , 𝑦 , 𝑧) = (π‘₯ , 𝑦 , 0) for all (π‘₯ , 𝑦 , 𝑧) ∈ 𝐅 3 , which illustrates the equation 𝑇𝑇 † = 𝑃 range 𝑇 from 6.69(b). Annotated Entity: ID: 237 Spans: True Boxes: True Text: 224 Chapter 6 Inner Product Spaces Exercises 6C 1 Suppose 𝑣 1 , … , 𝑣 π‘š ∈ 𝑉 . Prove that {𝑣 1 , … , 𝑣 π‘š } βŸ‚ = ( span (𝑣 1 , … , 𝑣 π‘š )) βŸ‚ . 2 Suppose π‘ˆ is a subspace of 𝑉 with basis 𝑒 1 , … , 𝑒 π‘š and 𝑒 1 , … , 𝑒 π‘š , 𝑣 1 , … , 𝑣 𝑛 is a basis of 𝑉 . Prove that if the Gram–Schmidt procedure is applied to the basis of 𝑉 above, producing a list 𝑒 1 , … , 𝑒 π‘š , 𝑓 1 , … , 𝑓 𝑛 , then 𝑒 1 , … , 𝑒 π‘š is an orthonormal basis of π‘ˆ and 𝑓 1 , … , 𝑓 𝑛 is an orthonormal basis of π‘ˆ βŸ‚ . 3 Suppose π‘ˆ is the subspace of 𝐑 4 defined by π‘ˆ = span ((1 , 2 , 3 , βˆ’4) , (βˆ’5 , 4 , 3 , 2)). Find an orthonormal basis of π‘ˆ and an orthonormal basis of π‘ˆ βŸ‚ . 4 Suppose 𝑒 1 , … , 𝑒 𝑛 is a list of vectors in 𝑉 with ‖𝑒 π‘˜ β€– = 1 for each π‘˜ = 1 , … , 𝑛 and ‖𝑣‖ 2 = βˆ£βŸ¨π‘£ , 𝑒 1 ⟩∣ 2 + β‹― + βˆ£βŸ¨π‘£ , 𝑒 𝑛 ⟩∣ 2 for all 𝑣 ∈ 𝑉 . Prove that 𝑒 1 , … , 𝑒 𝑛 is an orthonormal basis of 𝑉 . This exercise provides a converse to 6.30 ( b ) . 5 Suppose that 𝑉 is finite-dimensional and π‘ˆ is a subspace of 𝑉 . Show that 𝑃 π‘ˆ βŸ‚ = 𝐼 βˆ’ 𝑃 π‘ˆ , where 𝐼 is the identity operator on 𝑉 . 6 Suppose 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . Show that 𝑇 = 𝑇𝑃 ( null 𝑇) βŸ‚ = 𝑃 range 𝑇 𝑇. 7 Suppose that 𝑋 and π‘Œ are finite-dimensional subspaces of 𝑉 . Prove that 𝑃 𝑋 𝑃 π‘Œ = 0 if and only if ⟨π‘₯ , π‘¦βŸ© = 0 for all π‘₯ ∈ 𝑋 and all 𝑦 ∈ π‘Œ . 8 Suppose π‘ˆ is a finite-dimensional subspace of 𝑉 and 𝑣 ∈ 𝑉 . Define a linear functional πœ‘ ∢ π‘ˆ β†’ 𝐅 by πœ‘(𝑒) = βŸ¨π‘’ , π‘£βŸ© for all 𝑒 ∈ π‘ˆ . By the Riesz representation theorem (6.42) as applied to the inner product space π‘ˆ , there exists a unique vector 𝑀 ∈ π‘ˆ such that πœ‘(𝑒) = βŸ¨π‘’ , π‘€βŸ© for all 𝑒 ∈ π‘ˆ . Show that 𝑀 = 𝑃 π‘ˆ 𝑣 . 9 Suppose 𝑉 is finite-dimensional. Suppose 𝑃 ∈ β„’ (𝑉) is such that 𝑃 2 = 𝑃 and every vector in null 𝑃 is orthogonal to every vector in range 𝑃 . Prove that there exists a subspace π‘ˆ of 𝑉 such that 𝑃 = 𝑃 π‘ˆ . Annotated Entity: ID: 238 Spans: True Boxes: True Text: Section 6C Orthogonal Complements and Minimization Problems 225 10 Suppose 𝑉 is finite-dimensional and 𝑃 ∈ β„’ (𝑉) is such that 𝑃 2 = 𝑃 and ‖𝑃𝑣‖ ≀ ‖𝑣‖ for every 𝑣 ∈ 𝑉 . Prove that there exists a subspace π‘ˆ of 𝑉 such that 𝑃 = 𝑃 π‘ˆ . 11 Suppose 𝑇 ∈ β„’ (𝑉) and π‘ˆ is a finite-dimensional subspace of 𝑉 . Prove that π‘ˆ is invariant under 𝑇 ⟺ 𝑃 π‘ˆ 𝑇𝑃 π‘ˆ = 𝑇𝑃 π‘ˆ . 12 Suppose 𝑉 is finite-dimensional, 𝑇 ∈ β„’ (𝑉) , and π‘ˆ is a subspace of 𝑉 . Prove that π‘ˆ and π‘ˆ βŸ‚ are both invariant under 𝑇 ⟺ 𝑃 π‘ˆ 𝑇 = 𝑇𝑃 π‘ˆ . 13 Suppose 𝐅 = 𝐑 and 𝑉 is finite-dimensional. For each 𝑣 ∈ 𝑉 , let πœ‘ 𝑣 denote the linear functional on 𝑉 defined by πœ‘ 𝑣 (𝑒) = βŸ¨π‘’ , π‘£βŸ© for all 𝑒 ∈ 𝑉 . (a) Show that 𝑣 ↦ πœ‘ 𝑣 is an injective linear map from 𝑉 to 𝑉 β€² . (b) Use (a) and a dimension-counting argument to show that 𝑣 ↦ πœ‘ 𝑣 is an isomorphism from 𝑉 onto 𝑉 β€² . The purpose of this exercise is to give an alternative proof of the Riesz representation theorem ( 6.42 and 6.58 ) when 𝐅 = 𝐑 . Thus you should not use the Riesz representation theorem as a tool in your solution. 14 Suppose that 𝑒 1 , … , 𝑒 𝑛 is an orthonormal basis of 𝑉 . Explain why the dual basis (see 3.112) of 𝑒 1 , … , 𝑒 𝑛 is 𝑒 1 , … , 𝑒 𝑛 under the identification of 𝑉 β€² with 𝑉 provided by the Riesz representation theorem (6.58). 15 In 𝐑 4 , let π‘ˆ = span ((1 , 1 , 0 , 0) , (1 , 1 , 1 , 2)). Find 𝑒 ∈ π‘ˆ such that ‖𝑒 βˆ’ (1 , 2 , 3 , 4)β€– is as small as possible. 16 Suppose 𝐢[βˆ’1 , 1] is the vector space of continuous real-valued functions on the interval [βˆ’1 , 1] with inner product given by ⟨ 𝑓 , π‘”βŸ© = ∫ 1 βˆ’1 𝑓 𝑔 for all 𝑓 , 𝑔 ∈ 𝐢[βˆ’1 , 1] . Let π‘ˆ be the subspace of 𝐢[βˆ’1 , 1] defined by π‘ˆ = { 𝑓 ∈ 𝐢[βˆ’1 , 1] ∢ 𝑓 (0) = 0}. (a) Show that π‘ˆ βŸ‚ = {0} . (b) Show that 6.49 and 6.52 do not hold without the finite-dimensional hypothesis. Annotated Entity: ID: 239 Spans: True Boxes: True Text: 226 Chapter 6 Inner Product Spaces 17 Find 𝑝 ∈ 𝒫 3 (𝐑) such that 𝑝(0) = 0 , 𝑝 β€² (0) = 0 , and ∫ 1 0 ∣2 + 3π‘₯ βˆ’ 𝑝(π‘₯)∣ 2 𝑑π‘₯ is as small as possible. 18 Find 𝑝 ∈ 𝒫 5 (𝐑) that makes ∫ πœ‹ βˆ’πœ‹ ∣ sin π‘₯ βˆ’ 𝑝(π‘₯)∣ 2 𝑑π‘₯ as small as possible. The polynomial 6.65 is an excellent approximation to the answer to this exercise, but here you are asked to find the exact solution, which involves powers of πœ‹ . A computer that can perform symbolic integration should help. 19 Suppose 𝑉 is finite-dimensional and 𝑃 ∈ β„’ (𝑉) is an orthogonal projection of 𝑉 onto some subspace of 𝑉 . Prove that 𝑃 † = 𝑃 . 20 Suppose 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . Show that null 𝑇 † = ( range 𝑇) βŸ‚ and range 𝑇 † = ( null 𝑇) βŸ‚ . 21 Suppose 𝑇 ∈ β„’ (𝐅 3 , 𝐅 2 ) is defined by 𝑇(π‘Ž , 𝑏 , 𝑐) = (π‘Ž + 𝑏 + 𝑐 , 2𝑏 + 3𝑐). (a) For (π‘₯ , 𝑦) ∈ 𝐅 2 , find a formula for 𝑇 † (π‘₯ , 𝑦) . (b) Verify that the equation 𝑇𝑇 † = 𝑃 range 𝑇 from 6.69(b) holds with the formula for 𝑇 † obtained in (a). (c) Verify that the equation 𝑇 † 𝑇 = 𝑃 ( null 𝑇) βŸ‚ from 6.69(c) holds with the formula for 𝑇 † obtained in (a). 22 Suppose 𝑉 is finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . Prove that 𝑇𝑇 † 𝑇 = 𝑇 and 𝑇 † 𝑇𝑇 † = 𝑇 † . Both formulas above clearly hold if 𝑇 is invertible because in that case we can replace 𝑇 † with 𝑇 βˆ’1 . 23 Suppose 𝑉 and π‘Š are finite-dimensional and 𝑇 ∈ β„’ (𝑉 , π‘Š) . Prove that (𝑇 † ) † = 𝑇. The equation above is analogous to the equation (𝑇 βˆ’1 ) βˆ’1 = 𝑇 that holds if 𝑇 is invertible.