Transpose & Dot Product Extended Example

Transpose & Dot Product Extended Example

Transpose & Dot Product Def: The transpose of an m × n matrix A is the n × m matrix AT whose columns are the rows of A. So: The columns of AT are the rows of A. The rows of AT are the columns of A. 21 43 1 2 3 Example: If A = , then AT = 2 5 : 4 5 6 4 5 3 6 Convention: From now on, vectors v 2 Rn will be regarded as \columns" (i.e.: n × 1 matrices). Therefore, vT is a \row vector" (a 1 × n matrix). Observation: Let v; w 2 Rn. Then vT w = v · w. This is because: 2 3 w1 T . v w = v1 ··· vn 4 . 5 = v1w1 + ··· + vnwn = v · w: wn Where theory is concerned, the key property of transposes is the following: Prop 18.2: Let A be an m × n matrix. Then for x 2 Rn and y 2 Rm: (Ax) · y = x · (AT y): Here, · is the dot product of vectors. Extended Example Let A be a 5 × 3 matrix, so A: R3 ! R5. ◦ N(A) is a subspace of ◦ C(A) is a subspace of The transpose AT is a matrix, so AT : ! ◦ C(AT ) is a subspace of ◦ N(AT ) is a subspace of Observation: Both C(AT ) and N(A) are subspaces of . Might there be a geometric relationship between the two? (No, they're not equal.) Hm... Also: Both N(AT ) and C(A) are subspaces of . Might there be a geometric relationship between the two? (Again, they're not equal.) Hm... Orthogonal Complements Def: Let V ⊂ Rn be a subspace. The orthogonal complement of V is the set ? n V = fx 2 R j x · v = 0 for every v 2 V g: So, V ? consists of the vectors which are orthogonal to every vector in V . Fact: If V ⊂ Rn is a subspace, then V ? ⊂ Rn is a subspace. Examples in R3: ◦ The orthogonal complement of V = f0g is V ? = R3 ◦ The orthogonal complement of V = fz-axisg is V ? = fxy-planeg ◦ The orthogonal complement of V = fxy-planeg is V ? = fz-axisg ◦ The orthogonal complement of V = R3 is V ? = f0g Examples in R4: ◦ The orthogonal complement of V = f0g is V ? = R4 ◦ The orthogonal complement of V = fw-axisg is V ? = fxyz-spaceg ◦ The orthogonal complement of V = fzw-planeg is V ? = fxy-planeg ◦ The orthogonal complement of V = fxyz-spaceg is V ? = fw-axisg ◦ The orthogonal complement of V = R4 is V ? = f0g Prop 19.3-19.4-19.5: Let V ⊂ Rn be a subspace. Then: (a) dim(V ) + dim(V ?) = n (b) (V ?)? = V (c) V \ V ? = f0g (d) V + V ? = Rn. Part (d) means: \Every vector x 2 Rn can be written as a sum x = v + w where v 2 V and w 2 V ?." Also, it turns out that the expression x = v + w is unique: that is, there is only one way to write x as a sum of a vector in V and a vector in V ?. Meaning of C(AT ) and N(AT ) Q: What does C(AT ) mean? Well, the columns of AT are the rows of A. So: C(AT ) = column space of AT = span of columns of AT = span of rows of A: For this reason: We call C(AT ) the row space of A. Q: What does N(AT ) mean? Well: x 2 N(AT ) () AT x = 0 () (AT x)T = 0T () xT A = 0T : So, for an m × n matrix A, we see that: N(AT ) = fx 2 Rm j xT A = 0T g. For this reason: We call N(AT ) the left null space of A. Relationships among the Subspaces Theorem: Let A be an m × n matrix. Then: ◦ C(AT ) = N(A)? ◦ N(AT ) = C(A)? Corollary: Let A be an m × n matrix. Then: ◦ C(A) = N(AT )? ◦ N(A) = C(AT )? Prop 18.3: Let A be an m × n matrix. Then rank(A) = rank(AT ). Motivating Questions for Reading Problem 1: Let b 2 C(A). So, the system of equations Ax = b does have solutions, possibly infinitely many. Q: What is the solution x of Ax = b with kxk the smallest? Problem 2: Let b 2= C(A). So, the system of equations Ax = b does not have any solutions. In other words, Ax − b 6= 0. Q: What is the vector x that minimizes the error kAx−bk? That is, what is the vector x that comes closest to being a solution to Ax = b? Orthogonal Projection Def: Let V ⊂ Rn be a subspace. Then every vector x 2 Rn can be written uniquely as x = v + w; where v 2 V and w 2 V ?: n n The orthogonal projection onto V is the function ProjV : R ! R given by: ProjV (x) = v. (Note that ProjV ? (x) = w.) Prop 20.1: Let V ⊂ Rn be a subspace. Then: ProjV + ProjV ? = In: Of course, we already knew this: We have x = v+w = ProjV (x)+ProjV ? (x). n Formula: Let fv1;:::; vkg be a basis of V ⊂ R . Let A be the n × k matrix 2 3 A = 4v1 ··· vk5: Then: T −1 T ProjV = A(A A) A : (∗) Geometry Observations: Let V ⊂ Rn be a subspace, and x 2 Rn a vector. (1) The distance from x to V is: kProjV ? (x)k = kx − ProjV (x)k. (2) The vector in V that is closest to x is: ProjV (x). Derivation of (∗): Notice ProjV (x) is a vector in V = span(v1;:::; vk) = C(A) = Range(A), and k therefore ProjV (x) = Ay for some vector y 2 R . ? ? T Now notice that x − ProjV (x) = x − Ay is a vector in V = C(A) = N(A ), which means that AT (x − Ay) = 0, which means AT x = AT Ay. Now, it turns out that our matrix AT A is invertible (proof in L20), so we get y = (AT A)−1AT x. T −1 T Thus, ProjV (x) = Ay = A(A A) A x. ♦ Minimum Magnitude Solution Prop 19.6: Let b 2 C(A) (so Ax = b has solutions). Then there exists T exactly one vector x0 2 C(A ) with Ax0 = b. And: Among all solutions of Ax = b, the vector x0 has the smallest length. In other words: There is exactly one vector x0 in the row space of A which solves Ax = b { and this vector is the solution of smallest length. To Find x0: Start with any solution x of Ax = b. Then x0 = ProjC(AT )(x): Least Squares Approximation Idea: Suppose b 2= C(A). So, Ax = b has no solutions, so Ax − b 6= 0. We want to find the vector x∗ which minimizes the error kAx∗ − bk. That is, we want the vector x∗ for which Ax∗ is the closest vector in C(A) to b. In other words, we want the vector x∗ for which Ax∗ − b is orthogonal to C(A). So, Ax∗ − b 2 C(A)? = N(AT ), meaning that AT (Ax∗ − b) = 0, i.e.: AT Ax∗ = AT b: Quadratic Forms (Intro) Given an m × n matrix A, we can regard it as a linear transformation T : Rn ! Rm. In the special case where the matrix A is a symmetric matrix, we can also regard A as defining a \quadratic form": Def: Let A be a symmetric n × n matrix. The quadratic form associated n to A is the function QA : R ! R given by: QA(x) = x · Ax (· is the dot product) 2 3 x1 T . = x Ax = x1 ··· xn A 4 . 5 xn Notice that quadratic forms are not linear transformations! Orthonormal Bases Def: A basis fw1;:::; wkg for a subspace V is an orthonormal basis if: (1) The basis vectors are mutually orthogonal: wi · wj = 0 (for i 6= j); (2) The basis vectors are unit vectors: wi · wi = 1. (i.e.: kwik = 1) Orthonormal bases are nice for (at least) two reasons: (a) It is much easier to find the B-coordinates [v]B of a vector when the basis B is orthonormal; (b) It is much easier to find the projection matrix onto a subspace V when we have an orthonormal basis for V . n Prop: Let fw1;:::; wkg be an orthonormal basis for a subspace V ⊂ R . (a) Every vector v 2 V can be written v = (v · w1)w1 + ··· + (v · wk)wk: (b) For all x 2 Rn: ProjV (x) = (x · w1)w1 + ··· + (x · wk)wk: T (c) Let A be the matrix with columns fw1;:::; wkg. Then A A = Ik, so: T −1 T T ProjV = A(A A) A = AA : Orthogonal Matrices Def: An orthogonal matrix is an invertible matrix C such that C−1 = CT : n Example: Let fv1;:::; vng be an orthonormal basis for R . Then the matrix 2 3 C = 4v1 ··· vn5 is an orthogonal matrix. In fact, every orthogonal matrix C looks like this: the columns of any orthogonal matrix form an orthonormal basis of Rn. Where theory is concerned, the key property of orthogonal matrices is: Prop 22.4: Let C be an orthogonal matrix. Then for v; w 2 Rn: Cv · Cw = v · w: Gram-Schmidt Process Since orthonormal bases have so many nice properties, it would be great if we had a way of actually manufacturing orthonormal bases. That is: n Goal: We are given a basis fv1;:::; vkg for a subspace V ⊂ R . We would like an orthonormal basis fw1;:::; wkg for our subspace V . Notation: We will let V1 = span(v1) V2 = span(v1; v2) . Vk = span(v1;:::; vk) = V: Idea: Build an orthonormal basis for V1, then for V2, :::; up to Vk = V .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    13 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us