Lecture 25: 6.3 Orthonormal Bases

Total Page:16

File Type:pdf, Size:1020Kb

Lecture 25: 6.3 Orthonormal Bases Lecture 25: 6.3 Orthonormal Bases Wei-Ta Chu 2008/12/24 Theorem 6.3.2 Theorem 6.3.2 If S is an orthonormal basis for an n-dimensional inner product space, and if (u)s = (u1, u2, …, un) and (v)s = (v1, v2, …, vn) then: 2 2 2 u u1 u2 un 2 2 2 d(u, v) (u1 v1) (u2 v2 ) (un vn ) u, v u1v1 u2v2 unvn Remark By working with orthonormal bases, the computation of general norms and inner products can be reduced to the computation of Euclidean norms and inner products of the coordinate vectors. 2008/12/24 Elementary Linear Algebra 2 Example If R3 has the Euclidean inner product, then the norm of the vector u=(1,1,1) is However, if we let R3 have the orthonormal basis S in the last example, then we know from that the coordinate vector of u relative to S is (u)s= (1, -1/5, 7/5) The norm of u yields 2008/12/24 Elementary Linear Algebra 3 Coordinates Relative to Orthogonal Bases If S = {v1, v2, …, vn} is an orthogonal basis for a vector space V, then normalizing each of these vectors yields the orthonormal basis v v v S' 1 , 2 ,, n v1 v2 vn Thus, if u is any vector in V, it follows from theorem 6.3.1 that v v v v v v u u, 1 1 u, 2 2 u, n n v1 v1 v2 v2 v n vn or u, v1 u, v2 u, vn u 2 v1 2 v2 2 vn v1 v 2 vn The above equation expresses u as a linear combination of the vectors in the orthogonal basis S. 2008/12/24 Elementary Linear Algebra 4 Theorem 6.3.3 Theorem 6.3.3 If S = {v1, v2, …, vn} is an orthogonal set of nonzero vectors in an inner product space, then S is linearly independent. 2008/12/24 Elementary Linear Algebra 5 Proof of Theorem 6.3.3 Assume that k1v1+k2v2+…+knvn = 0. To demonstrate that S is linearly independent, we must prove that k1=k2=…=0. For each vi in S, k1v1+k2v2+…+knvn, vi= 0,vi=0 or, equivalently k1v1,vi+ k2v2,vi+…+knvn,vi=0 From the orthogonality of S it follows that vj,vi=0 when j is not equal to i, so the equation reduces to kivi,vi=0 Since the vectors in S are assumed to be nonzero, vi,vi ≠0. Therefore, ki=0. Since the subscript i is arbitrary, we have k1=k2=…=kn=0. 2008/12/24 Elementary Linear Algebra 6 Theorem 6.3.4 Theorem 6.3.4 (Projection Theorem) If W is a finite-dimensional subspace of an inner product space V, then every vector u in V can be expressed in exactly one way as u = w1 + w2 where w1 is in W and w2 is in W . u u w2 w2 W w1 w1 W 2008/12/24 Elementary Linear Algebra 7 Projection u w2 W w1 The vector w1 is called the orthogonal projection of u on W and is denoted projWu. The vector w2 is called the component of u orthogonal to W and is denote by projWu. u = projWu + projWu Since w2 = u-w1, it follows that projWu = u –projWu So we can write u = projWu + (u –projWu) u (u –proj u) W W projWu 2008/12/24 Elementary Linear Algebra 8 Theorem 6.3.5 Theorem 6.3.5 Let W be a finite-dimensional subspace of an inner product space V. If {v1, …, vr} is an orthonormal basis for W, and u is any vector in V, then projwu = u,v1v1 + u,v2v2 + … + u,vrvr If {v1, …, vr} is an orthogonal basis for W, and u is any vector in V, then u, v1 u, v2 u, vr projW u 2 v1 2 v2 2 vr Need Normalization v1 v2 vr 2008/12/24 Elementary Linear Algebra 9 Example Let R3 have the Euclidean inner product, and let W be the subspace spanned by the orthonormal vectors v1 = (0, 1, 0) and v2 = (-4/5, 0, 3/5). From the above theorem, the orthogonal projection of u = (1, 1, 1) on W is projwu =<, u v 1 > v 1 <, u v 2 > v 2 1 4 3 4 3 =(1)(0, 1, 0) ( )( , 0, )=( , 1, ) 5 5 5 25 25 The component of u orthogonal to W is 4 3 21 28 proju = u proj u =(1,1,1) ( ,1, ) ( ,0, ) w w 25 25 25 25 Observe that projWu is orthogonal to both v1 and v2. 2008/12/24 Elementary Linear Algebra 10 Finding Orthogonal/Orthonormal Bases Theorem 6.3.6 Every nonzero finite-dimensional inner product space has an orthonormal basis. Remark The step-by-step construction for converting an arbitrary basis into an orthogonal basis is called the Gram-Schmidt process. 2008/12/24 Elementary Linear Algebra 11 Proof of Theorem 6.3.6 Let V be an nonzero finite-dimensional inner product space, and suppose that {u1, u2, …, un} is any basis for V. It suffices to show that V has an orthogonal basis, since the vectors in the orthogonal basis can be normalized to produce an orthonormal basis for V. The following sequence of steps will produce an orthogonal basis {v1, v2, …, vn} for V. Step 1: Let v1 = u1. 2008/12/24 Elementary Linear Algebra 12 Proof of Theorem 6.3.6 u2 v2 = u2 –projW1u2 W1 v1 projW1u2 Step 2: We can obtain a vector v2 that is orthogonal to v1 by computing the component of u2 that is orthogonal to the space W1 spanned by v1. Of course if v2=0, then v2 is not a basis vector. But this cannot happen. Assume v2=0, Which says that u2 is multiple of u1, contradicting the linear independence of the basis S={u1,u2,…,un} 2008/12/24 Elementary Linear Algebra 13 Proof of Theorem 6.3.6 Step 3: To construct a vector v3 that is orthogonal to both v1 and v2, we compute the component of u3 orthogonal to space W2 spanned by v1 and v2. From Theorem 6.3.5(b): As in the Step 2, the linear independence of {u1,u2,…,un} u ensures that v3 ≠ 0 3 v2 v1 W 2008/12/24 Elementary Linear Algebra 14 Proof of Theorem 6.3.6 Step 4: To demonstrate a vector v4 is orthogonal to v1, v2, and v3, we compute the component of u4 orthogonal to the space W3 spanned by v1, v2, and v3. Continuing in this way, we will obtain, after n steps, an orthogonal set of vectors {v1,v2,…,vn}. Since V is n- dimensional and every orthogonal set is linearly independence, the set {v1,v2,…,vn} is an orthogonal basis for V. 2008/12/24 Elementary Linear Algebra 15 Example (Gram-Schmidt Process) Consider the vector space R3 with the Euclidean inner product. Apply the Gram-Schmidt process to transform the basis vectors u1 = (1, 1, 1), u2 = (0, 1, 1), u3 = (0, 0, 1) into an orthogonal basis {v1, v2, v3}; then normalize the orthogonal basis vectors to obtain an orthonormal basis {q1, q2, q3}. Solution: Step 1: Let v1 = u1.That is, v1 = u1 = (1, 1, 1) Step 2: Let v = u –proj u . That is, 2 2 W1 2 u, v v uproj u u 2 1 v 2 2 w1 2 22 1 v1 2 2 1 1 (0, 1, 1) (1, 1, 1) ( , , ) 3 3 3 3 2008/12/24 Elementary Linear Algebra 16 Example (Gram-Schmidt Process) We have two vectors in W2 now! Step 3: Let v = u –proj u . That is, 3 3 W2 3 u,, v u v v uproj u u 3 1 v 3 2 v 3 3 w2 3 32 1 2 2 v1 v 2 1 1/ 3 2 1 1 1 1 (0, 1, 1) (1, 1, 1) ( , , ) (0, , ) 3 2 / 3 3 3 3 2 2 Thus, v1 = (1, 1, 1), v2 = (-2/3, 1/3, 1/3), v3 = (0, -1/2, 1/2) form an orthogonal basis for R3. The norms of these vectors are 6 1 v 3, v , v 1 23 3 2 so an orthonormal basis for R3 is v11 1 1 v 2 2 1 1 q1 ( , , ), q 2 ( , , ), v13 3 3 v 2 66 6 v3 1 1 q3 (0, - , ) v3 2 2 2008/12/24 Elementary Linear Algebra 17 Theorem 6.3.7 Theorem 6.3.7 (QR-Decomposition) If A is an mn matrix with linearly independent column vectors, then A can be factored as A = QR where Q is an mn matrix with orthonormal column vectors, and R is an nn invertible upper triangular matrix. Remark In recent years the QR-decomposition has assumed growing importance as the mathematical foundation for a wide variety of practical algorithms, including a widely used algorithm for computing eigenvalues of large matrices. 2008/12/24 Elementary Linear Algebra 18 QR-Decomposition Suppose that the column vectors of A are u1, u2, …, un and the orthonormal column vectors of Q are q1, q2, …, qn; thus A = [u1 | u2 | … | un] and Q = [q1 | q2 | … | qn] From Theorem 6.3.1, the vectors u1, u2, …, un are expressible in terms of the vectors q1, q2, …, qn … … … … 2008/12/24 Elementary Linear Algebra 19 QR-Decomposition Recalling from Section 1.3 that the jth column vector of a matrix product is a linear combination of the column vectors of the first factor with coefficients coming from the jth column of the second factor.
Recommended publications
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Orthogonal Bases and the -So in Section 4.8 We Discussed the Problem of Finding the Orthogonal Projection P
    Orthogonal Bases and the -So In Section 4.8 we discussed the problemR. of finding the orthogonal projection p the vector b into the V of , . , suhspace the vectors , v2 ho If v1 v,, form a for V, and the in x n matrix A has these basis vectors as its column vectors. ilt the orthogonal projection p is given by p = Ax where x is the (unique) solution of the normal system ATAx = A7b. The formula for p takes an especially simple and attractive Form when the ba vectors , . .. , v1 v are mutually orthogonal. DEFINITION Orthogonal Basis An orthogonal basis for the suhspacc V of R” is a basis consisting of vectors , ,v,, that are mutually orthogonal, so that v v = 0 if I j. If in additii these basis vectors are unit vectors, so that v1 . = I for i = 1. 2 n, thct the orthogonal basis is called an orthonormal basis. Example 1 The vectors = (1, 1,0), v2 = (1, —1,2), v3 = (—1,1,1) form an orthogonal basis for . We “normalize” ‘ R3 can this orthogonal basis 1w viding each basis vector by its length: If w1=—- (1=1,2,3), lvii 4.9 Orthogonal Bases and the Gram-Schmidt Algorithm 295 then the vectors /1 I 1 /1 1 ‘\ 1 2” / I w1 0) W2 = —— _z) W3 for . form an orthonormal basis R3 , . ..., v, of the m x ii matrix A Now suppose that the column vectors v v2 form an orthogonal basis for the suhspacc V of R’. Then V}.VI 0 0 v2.v ..
    [Show full text]
  • Orthonormal Bases in Hilbert Space APPM 5440 Fall 2017 Applied Analysis
    Orthonormal Bases in Hilbert Space APPM 5440 Fall 2017 Applied Analysis Stephen Becker November 18, 2017(v4); v1 Nov 17 2014 and v2 Jan 21 2016 Supplementary notes to our textbook (Hunter and Nachtergaele). These notes generally follow Kreyszig’s book. The reason for these notes is that this is a simpler treatment that is easier to follow; the simplicity is because we generally do not consider uncountable nets, but rather only sequences (which are countable). I have tried to identify the theorems from both books; theorems/lemmas with numbers like 3.3-7 are from Kreyszig. Note: there are two versions of these notes, with and without proofs. The version without proofs will be distributed to the class initially, and the version with proofs will be distributed later (after any potential homework assignments, since some of the proofs maybe assigned as homework). This version does not have proofs. 1 Basic Definitions NOTE: Kreyszig defines inner products to be linear in the first term, and conjugate linear in the second term, so hαx, yi = αhx, yi = hx, αy¯ i. In contrast, Hunter and Nachtergaele define the inner product to be linear in the second term, and conjugate linear in the first term, so hαx,¯ yi = αhx, yi = hx, αyi. We will follow the convention of Hunter and Nachtergaele, so I have re-written the theorems from Kreyszig accordingly whenever the order is important. Of course when working with the real field, order is completely unimportant. Let X be an inner product space. Then we can define a norm on X by kxk = phx, xi, x ∈ X.
    [Show full text]
  • Lecture 4: April 8, 2021 1 Orthogonality and Orthonormality
    Mathematical Toolkit Spring 2021 Lecture 4: April 8, 2021 Lecturer: Avrim Blum (notes based on notes from Madhur Tulsiani) 1 Orthogonality and orthonormality Definition 1.1 Two vectors u, v in an inner product space are said to be orthogonal if hu, vi = 0. A set of vectors S ⊆ V is said to consist of mutually orthogonal vectors if hu, vi = 0 for all u 6= v, u, v 2 S. A set of S ⊆ V is said to be orthonormal if hu, vi = 0 for all u 6= v, u, v 2 S and kuk = 1 for all u 2 S. Proposition 1.2 A set S ⊆ V n f0V g consisting of mutually orthogonal vectors is linearly inde- pendent. Proposition 1.3 (Gram-Schmidt orthogonalization) Given a finite set fv1,..., vng of linearly independent vectors, there exists a set of orthonormal vectors fw1,..., wng such that Span (fw1,..., wng) = Span (fv1,..., vng) . Proof: By induction. The case with one vector is trivial. Given the statement for k vectors and orthonormal fw1,..., wkg such that Span (fw1,..., wkg) = Span (fv1,..., vkg) , define k u + u = v − hw , v i · w and w = k 1 . k+1 k+1 ∑ i k+1 i k+1 k k i=1 uk+1 We can now check that the set fw1,..., wk+1g satisfies the required conditions. Unit length is clear, so let’s check orthogonality: k uk+1, wj = vk+1, wj − ∑ hwi, vk+1i · wi, wj = vk+1, wj − wj, vk+1 = 0. i=1 Corollary 1.4 Every finite dimensional inner product space has an orthonormal basis.
    [Show full text]
  • Ch 5: ORTHOGONALITY
    Ch 5: ORTHOGONALITY 5.5 Orthonormal Sets 1. a set fv1; v2; : : : ; vng is an orthogonal set if hvi; vji = 0, 81 ≤ i 6= j; ≤ n (note that orthogonality only makes sense in an inner product space since you need to inner product to check hvi; vji = 0. n For example fe1; e2; : : : ; eng is an orthogonal set in R . 2. fv1; v2; : : : ; vng is an orthogonal set =) v1; v2; : : : ; vn are linearly independent. 3. a set fv1; v2; : : : ; vng is an orthonormal set if hvi; vji = 0 and jjvijj = 1, 81 ≤ i 6= j; ≤ n (i.e. orthogonal and unit length). n For example fe1; e2; : : : ; eng is an orthonormal set in R . 4. given an orthogonal basis for a vector space V , we can always find an orthonormal basis for V by dividing each vector by its length (see Example 2 and 3 page 256) n 5. a space with an orthonormal basis behaves like the euclidean space R with the standard basis (it is easier to work with basis that are orthonormal, just like it is n easier to work with the standard basis versus other bases for R ) 6. if v is a vector in a inner product space V with fu1; u2; : : : ; ung an orthonormal basis, Pn Pn then we can write v = i=1 ciui = i=1hv; uiiui Pn 7. an easy way to find the inner product of two vectors: if u = i=1 aiui and v = Pn i=1 biui, where fu1; u2; : : : ; ung is an orthonormal basis for an inner product space Pn V , then hu; vi = i=1 aibi Pn 8.
    [Show full text]
  • Math 2331 – Linear Algebra 6.2 Orthogonal Sets
    6.2 Orthogonal Sets Math 2331 { Linear Algebra 6.2 Orthogonal Sets Jiwen He Department of Mathematics, University of Houston [email protected] math.uh.edu/∼jiwenhe/math2331 Jiwen He, University of Houston Math 2331, Linear Algebra 1 / 12 6.2 Orthogonal Sets Orthogonal Sets Basis Projection Orthonormal Matrix 6.2 Orthogonal Sets Orthogonal Sets: Examples Orthogonal Sets: Theorem Orthogonal Basis: Examples Orthogonal Basis: Theorem Orthogonal Projections Orthonormal Sets Orthonormal Matrix: Examples Orthonormal Matrix: Theorems Jiwen He, University of Houston Math 2331, Linear Algebra 2 / 12 6.2 Orthogonal Sets Orthogonal Sets Basis Projection Orthonormal Matrix Orthogonal Sets Orthogonal Sets n A set of vectors fu1; u2;:::; upg in R is called an orthogonal set if ui · uj = 0 whenever i 6= j. Example 82 3 2 3 2 39 < 1 1 0 = Is 4 −1 5 ; 4 1 5 ; 4 0 5 an orthogonal set? : 0 0 1 ; Solution: Label the vectors u1; u2; and u3 respectively. Then u1 · u2 = u1 · u3 = u2 · u3 = Therefore, fu1; u2; u3g is an orthogonal set. Jiwen He, University of Houston Math 2331, Linear Algebra 3 / 12 6.2 Orthogonal Sets Orthogonal Sets Basis Projection Orthonormal Matrix Orthogonal Sets: Theorem Theorem (4) Suppose S = fu1; u2;:::; upg is an orthogonal set of nonzero n vectors in R and W =spanfu1; u2;:::; upg. Then S is a linearly independent set and is therefore a basis for W . Partial Proof: Suppose c1u1 + c2u2 + ··· + cpup = 0 (c1u1 + c2u2 + ··· + cpup) · = 0· (c1u1) · u1 + (c2u2) · u1 + ··· + (cpup) · u1 = 0 c1 (u1 · u1) + c2 (u2 · u1) + ··· + cp (up · u1) = 0 c1 (u1 · u1) = 0 Since u1 6= 0, u1 · u1 > 0 which means c1 = : In a similar manner, c2,:::,cp can be shown to by all 0.
    [Show full text]
  • CLIFFORD ALGEBRAS Property, Then There Is a Unique Isomorphism (V ) (V ) Intertwining the Two Inclusions of V
    CHAPTER 2 Clifford algebras 1. Exterior algebras 1.1. Definition. For any vector space V over a field K, let T (V ) = k k k Z T (V ) be the tensor algebra, with T (V ) = V V the k-fold tensor∈ product. The quotient of T (V ) by the two-sided⊗···⊗ ideal (V ) generated byL all v w + w v is the exterior algebra, denoted (V ).I The product in (V ) is usually⊗ denoted⊗ α α , although we will frequently∧ omit the wedge ∧ 1 ∧ 2 sign and just write α1α2. Since (V ) is a graded ideal, the exterior algebra inherits a grading I (V )= k(V ) ∧ ∧ k Z M∈ where k(V ) is the image of T k(V ) under the quotient map. Clearly, 0(V )∧ = K and 1(V ) = V so that we can think of V as a subspace of ∧(V ). We may thus∧ think of (V ) as the associative algebra linearly gener- ated∧ by V , subject to the relations∧ vw + wv = 0. We will write φ = k if φ k(V ). The exterior algebra is commutative | | ∈∧ (in the graded sense). That is, for φ k1 (V ) and φ k2 (V ), 1 ∈∧ 2 ∈∧ [φ , φ ] := φ φ + ( 1)k1k2 φ φ = 0. 1 2 1 2 − 2 1 k If V has finite dimension, with basis e1,...,en, the space (V ) has basis ∧ e = e e I i1 · · · ik for all ordered subsets I = i1,...,ik of 1,...,n . (If k = 0, we put { } k { n } e = 1.) In particular, we see that dim (V )= k , and ∅ ∧ n n dim (V )= = 2n.
    [Show full text]
  • Geometric (Clifford) Algebra and Its Applications
    Geometric (Clifford) algebra and its applications Douglas Lundholm F01, KTH January 23, 2006∗ Abstract In this Master of Science Thesis I introduce geometric algebra both from the traditional geometric setting of vector spaces, and also from a more combinatorial view which simplifies common relations and opera- tions. This view enables us to define Clifford algebras with scalars in arbitrary rings and provides new suggestions for an infinite-dimensional approach. Furthermore, I give a quick review of classic results regarding geo- metric algebras, such as their classification in terms of matrix algebras, the connection to orthogonal and Spin groups, and their representation theory. A number of lower-dimensional examples are worked out in a sys- tematic way using so called norm functions, while general applications of representation theory include normed division algebras and vector fields on spheres. I also consider examples in relativistic physics, where reformulations in terms of geometric algebra give rise to both computational and conceptual simplifications. arXiv:math/0605280v1 [math.RA] 10 May 2006 ∗Corrected May 2, 2006. Contents 1 Introduction 1 2 Foundations 3 2.1 Geometric algebra ( , q)...................... 3 2.2 Combinatorial CliffordG V algebra l(X,R,r)............. 6 2.3 Standardoperations .........................C 9 2.4 Vectorspacegeometry . 13 2.5 Linearfunctions ........................... 16 2.6 Infinite-dimensional Clifford algebra . 19 3 Isomorphisms 23 4 Groups 28 4.1 Group actions on .......................... 28 4.2 TheLipschitzgroupG ......................... 30 4.3 PropertiesofPinandSpingroups . 31 4.4 Spinors ................................ 34 5 A study of lower-dimensional algebras 36 5.1 (R1) ................................. 36 G 5.2 (R0,1) =∼ C -Thecomplexnumbers . 36 5.3 G(R0,0,1)...............................
    [Show full text]
  • True False Questions from 4,5,6
    (c)2015 UM Math Dept licensed under a Creative Commons By-NC-SA 4.0 International License. Math 217: True False Practice Professor Karen Smith 1. For every 2 × 2 matrix A, there exists a 2 × 2 matrix B such that det(A + B) 6= det A + det B. Solution note: False! Take A to be the zero matrix for a counterexample. 2. If the change of basis matrix SA!B = ~e4 ~e3 ~e2 ~e1 , then the elements of A are the same as the element of B, but in a different order. Solution note: True. The matrix tells us that the first element of A is the fourth element of B, the second element of basis A is the third element of B, the third element of basis A is the second element of B, and the the fourth element of basis A is the first element of B. 6 6 3. Every linear transformation R ! R is an isomorphism. Solution note: False. The zero map is a a counter example. 4. If matrix B is obtained from A by swapping two rows and det A < det B, then A is invertible. Solution note: True: we only need to know det A 6= 0. But if det A = 0, then det B = − det A = 0, so det A < det B does not hold. 5. If two n × n matrices have the same determinant, they are similar. Solution note: False: The only matrix similar to the zero matrix is itself. But many 1 0 matrices besides the zero matrix have determinant zero, such as : 0 0 6.
    [Show full text]
  • C Self-Adjoint Operators and Complete Orthonormal Bases
    C Self-adjoint operators and complete orthonormal bases The concept of the adjoint of an operator plays a very important role in many aspects of linear algebra and functional analysis. Here we will primar­ ily focus on s e ll~adjoint operators and how they can be used to obtain and characterize complete orthonormal sets of vectors in a Hilbert space. The main fact we will present here is that self-adjoint operators typically have orthogonal eigenvectors, and under some conditions, these eigenvectors can form a complete orthonormal basis for the Hilbert space. In particular, we will use this technique to find a variety of orthonormal bases for L2 [a, b] that satisfy different types of boundary conditions. We begin with the algebraic aspects of self-adjoint operators and their eigenvectors. Those algebraic properties are identical in the finite and infinite dimensional cases. The completeness arguments in infinite dimensions are more delicate, we will only state those results without complete proofs. We begin first by defining the adjoint. D efinition C.l. (Finite dimensional or bounded operator case) Let A H I -; H2 be a bounded operator between the Hilbert spaces HI and H 2. The adjomt A' : H2 ---; HI of Xis defined by the req1Lirement J (V2, AVI ) ~ = (A''V2, v])], (C.l) fOT all VI E HI--- and'V2 E H2· Note that for emphasis, we have written (. , .)] and (. , ')2 for the inner products in HI and H2 respectively. It remains to be shown that the relation (C.l) defines an operator A' from A uniquely. We will not do this here in general but will illustrate it with several examples.
    [Show full text]
  • Orthogonal Bases
    Orthogonal bases • Recall: Suppose that v1 , , vn are nonzero and (pairwise) orthogonal. Then v1 , , vn are independent. Definition 1. A basis v1 , , vn of a vector space V is an orthogonal basis if the vectors are (pairwise) orthogonal. 1 1 0 3 Example 2. Are the vectors − 1 , 1 , 0 an orthogonal basis for R ? 0 0 1 Solution. Note that we do not need to check that the three vectors are independent. That follows from their orthogonality. Example 3. Suppose v1 , , vn is an orthogonal basis of V, and that w is in V. Find c1 , , cn such that w = c1 v1 + + cnvn. Solution. Take the dot product of v1 with both sides: If v1 , , vn is an orthogonal basis of V, and w is in V, then w · v j w = c1 v1 + + cnvn with cj = . v j · v j Armin Straub 1 [email protected] 3 1 1 0 Example 4. Express 7 in terms of the basis − 1 , 1 , 0 . 4 0 0 1 Solution. Definition 5. A basis v1 , , vn of a vector space V is an orthonormal basis if the vectors are orthogonal and have length 1 . If v1 , , vn is an orthonormal basis of V, and w is in V, then w = c1 v1 + + cnvn with cj = v j · w. 1 1 0 Example 6. Is the basis − 1 , 1 , 0 orthonormal? If not, normalize the vectors 0 0 1 to produce an orthonormal basis. Solution. Armin Straub 2 [email protected] Orthogonal projections Definition 7. The orthogonal projection of vector x onto vector y is x · y xˆ = y.
    [Show full text]
  • Worksheet and Solutions
    Tuesday, December 6 ∗∗ Orthogonal bases & Orthogonal projections ∗∗ Solutions 1. A basis B is called an orthonormal basis if it is orthogonal and each basis vector has norm equal to 1. (a) Convert the orthogonal basis 80− 1 0 1 0 19 < 1 1 1 = B = @ 1 A , @ 1 A , @1A : 0 −2 1 ; into an orthonormal basis C . Solution. We just scale each basis vector by its length. The new, orthonormal, basis is 8 0−11 0 1 1 0119 < 1 1 1 = C = p @ 1 A , p @ 1 A , p @1A 2 6 3 : 0 −2 1 ; 011 0−41 (b) Find the coordinates of the vectors v1 = @3A and v2 = @ 2 A in the basis C . 5 7 Solution. The formula is the same as for a general orthogonal basis: writing u1, u2, and u3 for the basis vectors, the formula for each coordinate of v1 in the basis C is u · v i 1 = · 2 ui v1. kuik We compute to find 2 p −6 p 9 p u1 · v1 = p = 2, u2 · v1 = p = − 6, u3 · v1 = p = 3 3, 2 6 3 0 p 1 p2 B C so (v1)C = @−p 6A. Similarly, we get 3 3 r 6 p −16 2 5 u1 · v2 = p = 3 2, u2 · v2 = p = −8 , u3 · v2 = p , 2 6 3 3 0 p 1 3p 2 B C so (v2)C = @−8 p2/3A. 5/ 3 Orthonormal bases are very convenient for calculations. 2. (The Gram-Schmidt process) There is a standard process for converting a basis into an 3 orthogonal basis.
    [Show full text]