Euclidean Space

Euclidean Space

CHAPTER 3 Euclidean Space n We define the geometric concepts of length, distance, angle and perpendicularity for R . This gives n R the structure of an Euclidean Space. 3.1 Inner Product n We write a point u 2 R as a column vector, i.e. 1 × n matrix. n Definition 3.1. The inner product of u; v 2 R , i.e. 0 1 0 1 u1 v1 B . C B . C u = @ . A ; v = @ . A un vn is given by 0 1 v1 n T B . C X u · v := u v = u1 ··· un @ . A = uivi 2 R i=1 vn Note. To avoid confusion, I will omit the dot for scalar multiplication: I use cu instead of c · u. Some easily checked properties: -17- Chapter 3. Euclidean Space 3.1. Inner Product n Theorem 3.2. Let u; v; w 2 R ; c 2 R (1) u · v = v · u (2)( u + v) · w = u · w + v · w (3)( cu) · v = c(u · v) = u · (cv) (4) u · u ≥ 0 and u · u = 0 iff u = 0. More generally: Definition 3.3. Any vector space V over R equipped with an inner product V ×V −! R satisfying n Theorem 3.2 is called an inner product space. When V = R it is called an Euclidean space. Example 3.1 (Optional). An example of inner product space that is infinite dimensional: Let C[a; b] be the vector space of real-valued continuous function defined on a closed interval [a; b] ⊂ R. Then for f; g 2 C[a; b], Z b f · g := f(t)g(t)dt a gives an inner product on C[a; b]. Remark. All the Definitions and Theorems below applies to inner product spaces. Remark. When K = C, the inner product involves the complex conjugate 0 1 v1 n ∗ B . C X u · v := u v := u1 ··· un @ . A = uivi 2 C i=1 vn so that the last property (4) can hold. Also the third property have to be replaced by (3*)( cu) · v = c(u · v) = u · (cv) Properties 3.4. If A = (aij) is an m × n matrix, then the matrix entries are given by 0 aij = ei · Aej n 0 m where fejg is the standard basis for R and feig is the standard basis for R . -18- Chapter 3. Euclidean Space 3.1. Inner Product Definition 3.5. The norm (or length) of v is the nonnegative scalar p q 2 2 kvk := v · v = v1 + ··· + vn 2 R≥0 For c 2 R, we have kcvk = jcjkvk. Definition 3.6. The vector u with unit length, i.e. kuk = 1 is called a unit vector. 1 Given v 6= 0, kvk v has unit length and is called the normalization of v 0 1 1 p B −2 C 4 p Example 3.2.v = B C 2 R has norm kvk = 12 + (−2)2 + 22 + 02 = 9 = 3. @ 2 A 0 0 1=3 1 1 B −2=3 C v = B C 3 @ 2=3 A 0 is a unit vector. n Definition 3.7. The distance between u; v 2 R is defined by dist(u; v) := ku − vk Theorem 3.8 (Law of cosine). The angle θ between u and v can be calculated by u · v = kukkvk cos θ When θ = 90◦, we have -19- Chapter 3. Euclidean Space 3.2. Orthogonal Basis n Definition 3.9. Two vectors u; v 2 R are orthogonal (or perpendicular) to each other if u · v = 0. Properties 3.10 (Pythagorean Theorem). If u · v = 0, ku + vk2 = kuk2 + kvk2 n Properties 3.11 (Cauchy-Schwarz Inequality). For all u; v 2 R , ju · vj ≤ kukkvk n Properties 3.12 (Triangle Inequality). For all u; v 2 R , ku + vk ≤ kuk + kvk 1 1 Example 3.3. and are orthogonal to each other in 2. 1 −1 R n Example 3.4.0 is orthogonal to every vector in R . 3.2 Orthogonal Basis n Definition 3.13. Let S = fu1; :::; upg 2 R . • S is called an orthogonal set if ui · uj = 0 for all i 6= j. n • If in addition S is a basis of W ⊂ R , it is called an orthogonal basis for W . • If in addition all vectors in S has unit norm, it is called an orthonormal basis for W . n Example 3.5. The standard basis fe1; e2; ··· eng for R is an orthonormal basis. 1 1 Example 3.6. The set ; is an orthogonal basis for 2. 1 −1 R ( p1 ! p1 !) Its rescaled version, the set 2 ; 2 is an orthonormal basis for 2. p1 − p1 R 2 2 -20- Chapter 3. Euclidean Space 3.3. Orthogonal Projection n Theorem 3.14. Let B = fu1; :::; upg be an orthogonal basis for a subspace W ⊂ R . Then for x 2 W we can solve for the coordinates with respect to B explicitly as 0 1 c1 B . C [x]B = @ . A cn where x · ui ci = ; i = 1; :::; p ui · ui 3.3 Orthogonal Projection n Definition 3.15. Let W ⊂ R be a subspace. The orthogonal complement of W is the set ? n W := fv 2 R : v · w = 0 for every w 2 W g Properties 3.16. We have the following properties: ? n • W is a subspace of R . • If L = W ?, then W = L?. • x 2 W ? iff x is orthogonal to every vector in a spanning set ot W . Theorem 3.17. Let A 2 Mm×n(R). Then (RowA)? = NulA; (ColA)? = NulAT Definition 3.18. The orthogonal projection of b onto u is given by b · u Proj (b) := (b · e)e = u u u · u u where e := kuk is the normalization of u. -21- Chapter 3. Euclidean Space 3.3. Orthogonal Projection n n Theorem 3.19 (Orthogonal Decomposition Theorem). Let W ⊂ R a subspace. Then each x 2 R can be written uniquely in the form x = xb + z ? where xb 2 W and z 2 W . Therefore we have n ? R = W ⊕ W n n We sometimes write ProjW (x) := xb. Note that ProjW 2 L(R ; R ) with ? Im(ProjW ) = W; Ker(ProjW ) = W Proof. Explicitly, if fu1; :::; upg is an orthogonal basis of W , then x = Proj (x) + ··· + Proj (x) b u1 up x · u1 x · up = u1 + ··· + up u1 · u1 up · up and z = x − xb. W ? z x Proj (x) u2 u2 u1 Proj (x) u1 W = Span(u1; u2) ProjW (x) Figure 3.1: Orthogonal Projection Remark. In particular, the uniqueness statement says that the orthogonal decomposition, i.e. the formula for xb, does not depend on the basis used for W in the proof. -22- Chapter 3. Euclidean Space 3.3. Orthogonal Projection Properties 3.20. If x 2 W , then ProjW (x) = x. By using orthonormal basis, we can represent ProjW as a matrix: n Theorem 3.21. If fu1; :::; upg is an orthonormal basis for W ⊂ R , then ProjW (x) = (x · u1)u1 + ··· (x · up)up Equivalently, if U = u1 ··· up is an n × p matrix, then T ProjW (x) = UU x The matrix P := UU T is an n × n matrix which is called an orthogonal projection matrix. Definition 3.22. A projection matrix is an n × n matrix such that P 2 = P It is an orthogonal projection matrix if in addition P T = P 011 0 1 1 Example 3.7. If W = Span(v1; v2) where v1 = @0A ; v2 = @ 1 A then fv1; v2g is an orthogonal 1 −1 basis for W : v1 · v2 = 0. The normalization 1 0 p1 1 0 p 1 v 2 v 3 1 B C 2 B p1 C = 0 ; = 3 kv1k @ 1 A kv2k @ A p − p1 2 3 is then an orthonormal basis for W . We have 0 p1 p1 1 2 3 U = B 0 p1 C @ 3 A p1 − p1 2 3 and therefore 0 1 1 1 p p ! 0 5 1 1 1 2 3 p1 0 p1 6 3 6 T B 0 p1 C 2 2 1 1 1 ProjW = UU = 3 1 1 1 = @ 3 3 3 A @ A p p − p 1 1 5 p1 − p1 3 3 3 − 2 3 6 3 6 -23- Chapter 3. Euclidean Space 3.4. Orthogonal Matrix n n Theorem 3.23 (Best Approximation Theorem). Let W be a subspace of R and x 2 R . Then kx − ProjW xk ≤ kx − vk; for any v 2 W i.e. ProjW x 2 W is the closest point in W to x. 3.4 Orthogonal Matrix Definition 3.24. A linear transformation T 2 L(V; W ) between inner product spaces is called an isometry if it preserves the inner product: (T u) · (T v) = u · v for any vector u; v 2 V . n m Theorem 3.25. If T 2 L(R ; R ) is a linear isometry which is represented by an m × n matrix n U, then for x; y 2 R : • U has orthonormal columns T • U U = Idn×n •k Uxk = kxk (i.e. it preserves length) • (Ux) · (Uy) = 0 iff x · y = 0 (i.e. it preserves right angle) Definition 3.26. If n = m, the square matrix U corresponding to a linear isometry is called an orthogonal matrix. It is invertible with U−1 = UT The set of n × n orthogonal matrices is denoted by O(n). -24- Chapter 3. Euclidean Space 3.5. Gram-Schmidt Process Properties 3.27. Orthogonal matrices satisfy the following \group properties": • Idn×n 2 O(n).

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    14 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us