October 16, 2018 1 Orthogonality and Orthonormality

Total Page:16

File Type:pdf, Size:1020Kb

October 16, 2018 1 Orthogonality and Orthonormality Mathematical Toolkit Autumn 2018 Lecture 5: October 16, 2018 Lecturer: Madhur Tulsiani 1 Orthogonality and orthonormality. Definition 1.1 Two vectors u, v in an inner product space are said to be orthogonal if hu, vi = 0. A set of vectors S ⊆ V is said to consist of mutually orthogonal vectors if hu, vi = 0 for all u 6= v, u, v 2 S. A set of S ⊆ V is said to be orthonormal if hu, vi = 0 for all u 6= v, u, v 2 S and kuk = 1 for all u 2 S. Proposition 1.2 A set S ⊆ V n f0V g consisting of mutually orthogonal vectors is linearly inde- pendent. Proposition 1.3 (Gram-Schmidt orthogonalization) Given a finite set fv1,..., vng of linearly independent vectors, there exists a set of orthonormal vectors fw1,..., wng such that Span (fw1,..., wng) = Span (fv1,..., vng) . Proof: By induction. The case with one vector is trivial. Given the statement for k vectors and orthonormal fw1,..., wkg such that Span (fw1,..., wkg) = Span (fv1,..., vkg) , define k u + u = v − hw , v i · w and w = k 1 . k+1 k+1 ∑ i k+1 i k+1 k k i=1 uk+1 It is easy to check that the set fw1,..., wk+1g satisfies the required conditions. Corollary 1.4 Every finite dimensional inner product space has an orthonormal basis. In fact, Hilbert spaces also have orthonormal bases (which are countable). The existence of a maximal orthonormal set of vectors can be proved by using Zorn’s lemma, similar to the proof of existence of a Hamel basis for a vector space. However, we still need to prove that a maximal orthonormal set is a basis. This follows because we define the basis 1 slightly differently for a Hilbert space: instead of allowing only finite linear combinations, we allow infinite ones. The correct way of saying this is that is we still think of the span as the set of all finite linear combinations, then we only need that for any v 2 V, we can get arbitrarily close to v using elements in the span (a converging sequence of finite sums can get arbitrarily close to it’s limit). Thus, we only need that the span is dense in the Hilbert space V. However, if the maximal orthonormal set is not dense, then it is possible to show that it cannot be maximal. Such a basis is known as a Hilbert basis. Let V be a finite dimensional inner product space and let fw1,..., wng be an orthonormal basis for V. Then for any v 2 V, there exist c1,..., cn 2 F such that v = ∑i ci · wi. The coef- ficients ci are often called Fourier coefficients. Using the orthonormality and the properties of the inner product, we get ci = hwi, vi. This can be used to prove the following Proposition 1.5 (Parseval’s identity) Let V be a finite dimensional inner product space and let fw1,..., wng be an orthonormal basis for V. Then, for any u, v 2 V n hu, vi = ∑ hwi, ui · hwi, vi . i=1 2 Adjoint of a linear transformation Definition 2.1 Let V, W be inner product spaces over the same field F and let j : V ! W be a linear transformation. A transformation j∗ : W ! V is called an adjoint of j if hw, j(v)i = hj∗(w), vi 8v 2 V, w 2 W . n n Example 2.2 Let V = W = C with the inner product hu, vi = ∑i=1 ui · vi. Let j : V ! V be represented by the matrix A. Then j∗ is represented by the matrix AT. Exercise 2.3 Let jleft : Fib ! Fib be the left shift operator as before, and let h f , gi for f , g 2 Fib ¥ f (n)g(n) ∗ be defined as h f , gi = ∑n=0 Cn for C > 4. Find jleft. We will prove that every linear transformation has a unique adjoint. However, we first need the following characterization of linear transformations from V to F. Proposition 2.4 (Riesz Representation Theorem) Let V be a finite-dimensional inner product space over F and let a : V ! F be a linear transformation. Then there exists a unique z 2 V such that a(v) = hz, vi 8v 2 V. We only prove the theorem here for finite-dimensional spaces. However, the theorem holds for any Hilbert space. 2 Proof: Let fw1,..., wng be an orthonormal basis for V. Then check that n z = ∑ a(wi) · wi i=1 must be the unique z satisfying the required property. This can be used to prove the following: Proposition 2.5 Let V, W be finite dimensional inner product spaces and let j : V ! W be a linear transformation. Then there exists a unique j∗ : W ! V, such that hw, j(v)i = hj∗(w), vi 8v 2 V, w 2 W . Proof: For each w 2 W, the map hw, j(·)i : V ! F is a linear transformation (check!) and hence there exists a unique zw 2 V satisfying hw, j(v)i = hzw, vi 8v 2 V. Consider the map a : W ! V defined as a(w) = zw. By definition of a, hw, j(v)i = ha(w), vi 8v 2 V, w 2 W . To check that a is linear, we note that 8v 2 V, 8w1, w2 2 W, ha(w1 + w2), vi = hw1 + w2, j(v)i = hw1, j(v)i + hw2, j(v)i = ha(w1), vi + ha(w2), vi , which implies a(w1 + w2) = a(w1) + a(w2) (why?) a(c · w) = c · a(w) follows similarly. Note that the above proof only requires the Riesz representation theorem (to define zw) and hence also works for Hilbert spaces. 3.
Recommended publications
  • 21. Orthonormal Bases
    21. Orthonormal Bases The canonical/standard basis 011 001 001 B C B C B C B0C B1C B0C e1 = B.C ; e2 = B.C ; : : : ; en = B.C B.C B.C B.C @.A @.A @.A 0 0 1 has many useful properties. • Each of the standard basis vectors has unit length: q p T jjeijj = ei ei = ei ei = 1: • The standard basis vectors are orthogonal (in other words, at right angles or perpendicular). T ei ej = ei ej = 0 when i 6= j This is summarized by ( 1 i = j eT e = δ = ; i j ij 0 i 6= j where δij is the Kronecker delta. Notice that the Kronecker delta gives the entries of the identity matrix. Given column vectors v and w, we have seen that the dot product v w is the same as the matrix multiplication vT w. This is the inner product on n T R . We can also form the outer product vw , which gives a square matrix. 1 The outer product on the standard basis vectors is interesting. Set T Π1 = e1e1 011 B C B0C = B.C 1 0 ::: 0 B.C @.A 0 01 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 0 . T Πn = enen 001 B C B0C = B.C 0 0 ::: 1 B.C @.A 1 00 0 ::: 01 B C B0 0 ::: 0C = B. .C B. .C @. .A 0 0 ::: 1 In short, Πi is the diagonal square matrix with a 1 in the ith diagonal position and zeros everywhere else.
    [Show full text]
  • Orthonormal Bases in Hilbert Space APPM 5440 Fall 2017 Applied Analysis
    Orthonormal Bases in Hilbert Space APPM 5440 Fall 2017 Applied Analysis Stephen Becker November 18, 2017(v4); v1 Nov 17 2014 and v2 Jan 21 2016 Supplementary notes to our textbook (Hunter and Nachtergaele). These notes generally follow Kreyszig’s book. The reason for these notes is that this is a simpler treatment that is easier to follow; the simplicity is because we generally do not consider uncountable nets, but rather only sequences (which are countable). I have tried to identify the theorems from both books; theorems/lemmas with numbers like 3.3-7 are from Kreyszig. Note: there are two versions of these notes, with and without proofs. The version without proofs will be distributed to the class initially, and the version with proofs will be distributed later (after any potential homework assignments, since some of the proofs maybe assigned as homework). This version does not have proofs. 1 Basic Definitions NOTE: Kreyszig defines inner products to be linear in the first term, and conjugate linear in the second term, so hαx, yi = αhx, yi = hx, αy¯ i. In contrast, Hunter and Nachtergaele define the inner product to be linear in the second term, and conjugate linear in the first term, so hαx,¯ yi = αhx, yi = hx, αyi. We will follow the convention of Hunter and Nachtergaele, so I have re-written the theorems from Kreyszig accordingly whenever the order is important. Of course when working with the real field, order is completely unimportant. Let X be an inner product space. Then we can define a norm on X by kxk = phx, xi, x ∈ X.
    [Show full text]
  • Lecture 4: April 8, 2021 1 Orthogonality and Orthonormality
    Mathematical Toolkit Spring 2021 Lecture 4: April 8, 2021 Lecturer: Avrim Blum (notes based on notes from Madhur Tulsiani) 1 Orthogonality and orthonormality Definition 1.1 Two vectors u, v in an inner product space are said to be orthogonal if hu, vi = 0. A set of vectors S ⊆ V is said to consist of mutually orthogonal vectors if hu, vi = 0 for all u 6= v, u, v 2 S. A set of S ⊆ V is said to be orthonormal if hu, vi = 0 for all u 6= v, u, v 2 S and kuk = 1 for all u 2 S. Proposition 1.2 A set S ⊆ V n f0V g consisting of mutually orthogonal vectors is linearly inde- pendent. Proposition 1.3 (Gram-Schmidt orthogonalization) Given a finite set fv1,..., vng of linearly independent vectors, there exists a set of orthonormal vectors fw1,..., wng such that Span (fw1,..., wng) = Span (fv1,..., vng) . Proof: By induction. The case with one vector is trivial. Given the statement for k vectors and orthonormal fw1,..., wkg such that Span (fw1,..., wkg) = Span (fv1,..., vkg) , define k u + u = v − hw , v i · w and w = k 1 . k+1 k+1 ∑ i k+1 i k+1 k k i=1 uk+1 We can now check that the set fw1,..., wk+1g satisfies the required conditions. Unit length is clear, so let’s check orthogonality: k uk+1, wj = vk+1, wj − ∑ hwi, vk+1i · wi, wj = vk+1, wj − wj, vk+1 = 0. i=1 Corollary 1.4 Every finite dimensional inner product space has an orthonormal basis.
    [Show full text]
  • Orthonormality and the Gram-Schmidt Process
    Unit 4, Section 2: The Gram-Schmidt Process Orthonormality and the Gram-Schmidt Process The basis (e1; e2; : : : ; en) n n for R (or C ) is considered the standard basis for the space because of its geometric properties under the standard inner product: 1. jjeijj = 1 for all i, and 2. ei, ej are orthogonal whenever i 6= j. With this idea in mind, we record the following definitions: Definitions 6.23/6.25. Let V be an inner product space. • A list of vectors in V is called orthonormal if each vector in the list has norm 1, and if each pair of distinct vectors is orthogonal. • A basis for V is called an orthonormal basis if the basis is an orthonormal list. Remark. If a list (v1; : : : ; vn) is orthonormal, then ( 0 if i 6= j hvi; vji = 1 if i = j: Example. The list (e1; e2; : : : ; en) n n forms an orthonormal basis for R =C under the standard inner products on those spaces. 2 Example. The standard basis for Mn(C) consists of n matrices eij, 1 ≤ i; j ≤ n, where eij is the n × n matrix with a 1 in the ij entry and 0s elsewhere. Under the standard inner product on Mn(C) this is an orthonormal basis for Mn(C): 1. heij; eiji: ∗ heij; eiji = tr (eijeij) = tr (ejieij) = tr (ejj) = 1: 1 Unit 4, Section 2: The Gram-Schmidt Process 2. heij; ekli, k 6= i or j 6= l: ∗ heij; ekli = tr (ekleij) = tr (elkeij) = tr (0) if k 6= i, or tr (elj) if k = i but l 6= j = 0: So every vector in the list has norm 1, and every distinct pair of vectors is orthogonal.
    [Show full text]
  • Ch 5: ORTHOGONALITY
    Ch 5: ORTHOGONALITY 5.5 Orthonormal Sets 1. a set fv1; v2; : : : ; vng is an orthogonal set if hvi; vji = 0, 81 ≤ i 6= j; ≤ n (note that orthogonality only makes sense in an inner product space since you need to inner product to check hvi; vji = 0. n For example fe1; e2; : : : ; eng is an orthogonal set in R . 2. fv1; v2; : : : ; vng is an orthogonal set =) v1; v2; : : : ; vn are linearly independent. 3. a set fv1; v2; : : : ; vng is an orthonormal set if hvi; vji = 0 and jjvijj = 1, 81 ≤ i 6= j; ≤ n (i.e. orthogonal and unit length). n For example fe1; e2; : : : ; eng is an orthonormal set in R . 4. given an orthogonal basis for a vector space V , we can always find an orthonormal basis for V by dividing each vector by its length (see Example 2 and 3 page 256) n 5. a space with an orthonormal basis behaves like the euclidean space R with the standard basis (it is easier to work with basis that are orthonormal, just like it is n easier to work with the standard basis versus other bases for R ) 6. if v is a vector in a inner product space V with fu1; u2; : : : ; ung an orthonormal basis, Pn Pn then we can write v = i=1 ciui = i=1hv; uiiui Pn 7. an easy way to find the inner product of two vectors: if u = i=1 aiui and v = Pn i=1 biui, where fu1; u2; : : : ; ung is an orthonormal basis for an inner product space Pn V , then hu; vi = i=1 aibi Pn 8.
    [Show full text]
  • Math 2331 – Linear Algebra 6.2 Orthogonal Sets
    6.2 Orthogonal Sets Math 2331 { Linear Algebra 6.2 Orthogonal Sets Jiwen He Department of Mathematics, University of Houston [email protected] math.uh.edu/∼jiwenhe/math2331 Jiwen He, University of Houston Math 2331, Linear Algebra 1 / 12 6.2 Orthogonal Sets Orthogonal Sets Basis Projection Orthonormal Matrix 6.2 Orthogonal Sets Orthogonal Sets: Examples Orthogonal Sets: Theorem Orthogonal Basis: Examples Orthogonal Basis: Theorem Orthogonal Projections Orthonormal Sets Orthonormal Matrix: Examples Orthonormal Matrix: Theorems Jiwen He, University of Houston Math 2331, Linear Algebra 2 / 12 6.2 Orthogonal Sets Orthogonal Sets Basis Projection Orthonormal Matrix Orthogonal Sets Orthogonal Sets n A set of vectors fu1; u2;:::; upg in R is called an orthogonal set if ui · uj = 0 whenever i 6= j. Example 82 3 2 3 2 39 < 1 1 0 = Is 4 −1 5 ; 4 1 5 ; 4 0 5 an orthogonal set? : 0 0 1 ; Solution: Label the vectors u1; u2; and u3 respectively. Then u1 · u2 = u1 · u3 = u2 · u3 = Therefore, fu1; u2; u3g is an orthogonal set. Jiwen He, University of Houston Math 2331, Linear Algebra 3 / 12 6.2 Orthogonal Sets Orthogonal Sets Basis Projection Orthonormal Matrix Orthogonal Sets: Theorem Theorem (4) Suppose S = fu1; u2;:::; upg is an orthogonal set of nonzero n vectors in R and W =spanfu1; u2;:::; upg. Then S is a linearly independent set and is therefore a basis for W . Partial Proof: Suppose c1u1 + c2u2 + ··· + cpup = 0 (c1u1 + c2u2 + ··· + cpup) · = 0· (c1u1) · u1 + (c2u2) · u1 + ··· + (cpup) · u1 = 0 c1 (u1 · u1) + c2 (u2 · u1) + ··· + cp (up · u1) = 0 c1 (u1 · u1) = 0 Since u1 6= 0, u1 · u1 > 0 which means c1 = : In a similar manner, c2,:::,cp can be shown to by all 0.
    [Show full text]
  • True False Questions from 4,5,6
    (c)2015 UM Math Dept licensed under a Creative Commons By-NC-SA 4.0 International License. Math 217: True False Practice Professor Karen Smith 1. For every 2 × 2 matrix A, there exists a 2 × 2 matrix B such that det(A + B) 6= det A + det B. Solution note: False! Take A to be the zero matrix for a counterexample. 2. If the change of basis matrix SA!B = ~e4 ~e3 ~e2 ~e1 , then the elements of A are the same as the element of B, but in a different order. Solution note: True. The matrix tells us that the first element of A is the fourth element of B, the second element of basis A is the third element of B, the third element of basis A is the second element of B, and the the fourth element of basis A is the first element of B. 6 6 3. Every linear transformation R ! R is an isomorphism. Solution note: False. The zero map is a a counter example. 4. If matrix B is obtained from A by swapping two rows and det A < det B, then A is invertible. Solution note: True: we only need to know det A 6= 0. But if det A = 0, then det B = − det A = 0, so det A < det B does not hold. 5. If two n × n matrices have the same determinant, they are similar. Solution note: False: The only matrix similar to the zero matrix is itself. But many 1 0 matrices besides the zero matrix have determinant zero, such as : 0 0 6.
    [Show full text]
  • C Self-Adjoint Operators and Complete Orthonormal Bases
    C Self-adjoint operators and complete orthonormal bases The concept of the adjoint of an operator plays a very important role in many aspects of linear algebra and functional analysis. Here we will primar­ ily focus on s e ll~adjoint operators and how they can be used to obtain and characterize complete orthonormal sets of vectors in a Hilbert space. The main fact we will present here is that self-adjoint operators typically have orthogonal eigenvectors, and under some conditions, these eigenvectors can form a complete orthonormal basis for the Hilbert space. In particular, we will use this technique to find a variety of orthonormal bases for L2 [a, b] that satisfy different types of boundary conditions. We begin with the algebraic aspects of self-adjoint operators and their eigenvectors. Those algebraic properties are identical in the finite and infinite dimensional cases. The completeness arguments in infinite dimensions are more delicate, we will only state those results without complete proofs. We begin first by defining the adjoint. D efinition C.l. (Finite dimensional or bounded operator case) Let A H I -; H2 be a bounded operator between the Hilbert spaces HI and H 2. The adjomt A' : H2 ---; HI of Xis defined by the req1Lirement J (V2, AVI ) ~ = (A''V2, v])], (C.l) fOT all VI E HI--- and'V2 E H2· Note that for emphasis, we have written (. , .)] and (. , ')2 for the inner products in HI and H2 respectively. It remains to be shown that the relation (C.l) defines an operator A' from A uniquely. We will not do this here in general but will illustrate it with several examples.
    [Show full text]
  • Worksheet and Solutions
    Tuesday, December 6 ∗∗ Orthogonal bases & Orthogonal projections ∗∗ Solutions 1. A basis B is called an orthonormal basis if it is orthogonal and each basis vector has norm equal to 1. (a) Convert the orthogonal basis 80− 1 0 1 0 19 < 1 1 1 = B = @ 1 A , @ 1 A , @1A : 0 −2 1 ; into an orthonormal basis C . Solution. We just scale each basis vector by its length. The new, orthonormal, basis is 8 0−11 0 1 1 0119 < 1 1 1 = C = p @ 1 A , p @ 1 A , p @1A 2 6 3 : 0 −2 1 ; 011 0−41 (b) Find the coordinates of the vectors v1 = @3A and v2 = @ 2 A in the basis C . 5 7 Solution. The formula is the same as for a general orthogonal basis: writing u1, u2, and u3 for the basis vectors, the formula for each coordinate of v1 in the basis C is u · v i 1 = · 2 ui v1. kuik We compute to find 2 p −6 p 9 p u1 · v1 = p = 2, u2 · v1 = p = − 6, u3 · v1 = p = 3 3, 2 6 3 0 p 1 p2 B C so (v1)C = @−p 6A. Similarly, we get 3 3 r 6 p −16 2 5 u1 · v2 = p = 3 2, u2 · v2 = p = −8 , u3 · v2 = p , 2 6 3 3 0 p 1 3p 2 B C so (v2)C = @−8 p2/3A. 5/ 3 Orthonormal bases are very convenient for calculations. 2. (The Gram-Schmidt process) There is a standard process for converting a basis into an 3 orthogonal basis.
    [Show full text]
  • Inner Product Spaces
    CHAPTER 6 Woman teaching geometry, from a fourteenth-century edition of Euclid’s geometry book. Inner Product Spaces In making the definition of a vector space, we generalized the linear structure (addition and scalar multiplication) of R2 and R3. We ignored other important features, such as the notions of length and angle. These ideas are embedded in the concept we now investigate, inner products. Our standing assumptions are as follows: 6.1 Notation F, V F denotes R or C. V denotes a vector space over F. LEARNING OBJECTIVES FOR THIS CHAPTER Cauchy–Schwarz Inequality Gram–Schmidt Procedure linear functionals on inner product spaces calculating minimum distance to a subspace Linear Algebra Done Right, third edition, by Sheldon Axler 164 CHAPTER 6 Inner Product Spaces 6.A Inner Products and Norms Inner Products To motivate the concept of inner prod- 2 3 x1 , x 2 uct, think of vectors in R and R as x arrows with initial point at the origin. x R2 R3 H L The length of a vector in or is called the norm of x, denoted x . 2 k k Thus for x .x1; x2/ R , we have The length of this vector x is p D2 2 2 x x1 x2 . p 2 2 x1 x2 . k k D C 3 C Similarly, if x .x1; x2; x3/ R , p 2D 2 2 2 then x x1 x2 x3 . k k D C C Even though we cannot draw pictures in higher dimensions, the gener- n n alization to R is obvious: we define the norm of x .x1; : : : ; xn/ R D 2 by p 2 2 x x1 xn : k k D C C The norm is not linear on Rn.
    [Show full text]
  • Geometric Algebra Techniques for General Relativity
    Geometric Algebra Techniques for General Relativity Matthew R. Francis∗ and Arthur Kosowsky† Dept. of Physics and Astronomy, Rutgers University 136 Frelinghuysen Road, Piscataway, NJ 08854 (Dated: February 4, 2008) Geometric (Clifford) algebra provides an efficient mathematical language for describing physical problems. We formulate general relativity in this language. The resulting formalism combines the efficiency of differential forms with the straightforwardness of coordinate methods. We focus our attention on orthonormal frames and the associated connection bivector, using them to find the Schwarzschild and Kerr solutions, along with a detailed exposition of the Petrov types for the Weyl tensor. PACS numbers: 02.40.-k; 04.20.Cv Keywords: General relativity; Clifford algebras; solution techniques I. INTRODUCTION Geometric (or Clifford) algebra provides a simple and natural language for describing geometric concepts, a point which has been argued persuasively by Hestenes [1] and Lounesto [2] among many others. Geometric algebra (GA) unifies many other mathematical formalisms describing specific aspects of geometry, including complex variables, matrix algebra, projective geometry, and differential geometry. Gravitation, which is usually viewed as a geometric theory, is a natural candidate for translation into the language of geometric algebra. This has been done for some aspects of gravitational theory; notably, Hestenes and Sobczyk have shown how geometric algebra greatly simplifies certain calculations involving the curvature tensor and provides techniques for classifying the Weyl tensor [3, 4]. Lasenby, Doran, and Gull [5] have also discussed gravitation using geometric algebra via a reformulation in terms of a gauge principle. In this paper, we formulate standard general relativity in terms of geometric algebra. A comprehensive overview like the one presented here has not previously appeared in the literature, although unpublished works of Hestenes and of Doran take significant steps in this direction.
    [Show full text]
  • Inner Product Spaces Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 2, 2007)
    MAT067 University of California, Davis Winter 2007 Inner Product Spaces Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 2, 2007) The abstract definition of vector spaces only takes into account algebraic properties for the addition and scalar multiplication of vectors. For vectors in Rn, for example, we also have geometric intuition which involves the length of vectors or angles between vectors. In this section we discuss inner product spaces, which are vector spaces with an inner product defined on them, which allow us to introduce the notion of length (or norm) of vectors and concepts such as orthogonality. 1 Inner product In this section V is a finite-dimensional, nonzero vector space over F. Definition 1. An inner product on V is a map ·, · : V × V → F (u, v) →u, v with the following properties: 1. Linearity in first slot: u + v, w = u, w + v, w for all u, v, w ∈ V and au, v = au, v; 2. Positivity: v, v≥0 for all v ∈ V ; 3. Positive definiteness: v, v =0ifandonlyifv =0; 4. Conjugate symmetry: u, v = v, u for all u, v ∈ V . Remark 1. Recall that every real number x ∈ R equals its complex conjugate. Hence for real vector spaces the condition about conjugate symmetry becomes symmetry. Definition 2. An inner product space is a vector space over F together with an inner product ·, ·. Copyright c 2007 by the authors. These lecture notes may be reproduced in their entirety for non- commercial purposes. 2NORMS 2 Example 1. V = Fn n u =(u1,...,un),v =(v1,...,vn) ∈ F Then n u, v = uivi.
    [Show full text]