Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics, which at times may have seemed disjointed and unrelated to each other. However, the point is that there is much common ground here and having it all laid out in one place should be useful. Most of this material should be good to look at when preparing for the midterm. Some of it takes a \higher" level view of some of these concepts which won't necessarily be tested on the midterm, but it doesn't hurt to understand this higher-level view as well. When in doubt, ask me if it's not clear whether something here is something you should know, but the answer will most likely be yes. Orthogonal Projections Let's start with the most important property of orthonormal bases: n Fact. If ~u1; : : : ; ~uk is an orthonormal basis of a subspace V of R , then any ~x in V can be written as ~x = (~x · ~u1)~u1 + ··· + (~x · ~uk)~uk: In other words, the coefficients needed to express ~x as a linear combination of ~u1; : : : ; ~uk are obtained by taking dot products of ~x with the ~u's themselves. This is a crucially important fact! Recall that usually to express a vector ~x as a linear combi- nation of vectors ~v1; : : : ;~vk requires setting up the equation 0 1 0 j j 1 c1 B . C ~x = c1~v1 + ··· + ck~vk or @v1 ··· vkA @ . A = ~x j j ck and using row operations to solve for c1; : : : ; ck. The point is that having an orthonormal set of vectors greatly simplifies this since no row operations are necessary, we only need to compute some dot products. For this to be useful requires that we be able to find orthonormal bases of any subspace we want, but of course this is precisely what the Gram-Schmidt process allows us to do: n Fact. The Gram-Schmidt process applied to a basis ~v1; : : : ;~vk of a subspace V of R produces an orthonormal basis ~u1; : : : ; ~uk of that same subspace. Now we're in business. For instance, we can now easily compute orthogonal projections: n Fact. If ~u1; : : : ; ~uk is an orthonormal basis of a subspace V of R , the orthogonal projection of a vector ~x onto V is projV ~x = (~x · ~u1)~u1 + ··· (~x · uk)~uk: This can also be written as 0 j j 1 T projV ~x = QQ ~x where Q = @~u1 ··· ~ukA : j j n T Thus the matrix of the orthogonal projection of R onto the subspace V is QQ where Q is the matrix having the given orthonormal basis vectors as its columns. Note that this matrix QQT is always an n × n matrix and is symmetric since (QQT )T = (QT )T QT = QQT using the fact that the transpose of a product of matrices is the product of the transposes of each in the reverse order. This is our first contact between things related to orthogonality and things related to symmetry. We'll come back to this. Note also that the right-hand side of the first expression in the Fact above is precisely what you get when you project ~x onto the line spanned by each basis vector and add the results together, so that the projection of ~x is obtained simply by adding together these individual projections. This only works when you have an orthogonal basis: for a basis which is orthogonal but not necessarily orthonormal all that changes is that the dot products on the right-hand side get replaced by fractions with denominators coming from the general formula for the projection of a vector onto a line. So, to sum up, computing orthogonal projections involves the following steps: Fact. To compute orthogonal projections, you 1. Find a basis of the space you're projecting onto. 2. Apply the Gram-Schmidt process to that basis to get an orthonormal basis 3. Use that orthonormal basis to compute the projection as in the first part of the previous Fact, or use that orthonormal basis to compute the matrix of the projection as in the second part of the previous Fact. Least Squares Now, why do we care about orthogonal projections? Here is a key point: n n Fact. Given a subspace V of R and a vector ~x in R , the vector in V \closest" to ~x is the projection projV ~x. Here \closest" means that k~x − ~vk as v ranges through all things in V is minimized when ~v = projV ~x. So, orthogonal projections come up in problems dealing with minimizing some quantity. Here is an example: Example. We want to find the vector in the plane V spanned by 011 011 @1A ; @0A 1 1 1 which is closest to −1 . According to the above fact, this vector should be the orthogonal −2 1 projection of −1 onto the given plane. To compute this projection, we first find an orthonormal −2 basis of this plane. Applying the Gram-Schmidt process to the two given spanning vectors produces 0 p 1 0 p 1 1=p3 1= p6 @1=p3A ; @−2=p 6A 1= 3 1= 6 2 1 as an orthonormal basis. Hence the orthogonal projection of −1 onto this plane (using a formula −2 for orthogonal projections given in a previous Fact) is: 0 1 20 1 0 p 13 0 p 1 20 1 0 p 13 0 p 1 1 1 1=p3 1=p3 1 1= p6 1= p6 projV @−1A = 4@−1A · @1=p3A5 @1=p3A + 4@−1A · @−2=p 6A5 @−2=p 6A −2 −2 1= 3 1= 3 −2 1= 6 1= 6 p p 01= 31 0 1= 6 1 2 p 1 p = −p @1=p3A + p @−2=p 6A 3 1= 3 6 1= 6 0−2=31 0 1=6 1 = @−2=3A + @−2=6A −2=3 1=6 0−1=21 = @ −1 A : −1=2 1 If we want, the distance between this and −1 is −2 0 1 0 1 0 1 −1=2 1 −3=2 p 18 @ −1 A − @−1A = @ 0 A = : 2 −1=2 −2 3=2 Now, admittedly, this took quite a bit of work, but notice that there is another way to get this answer. Let A be the matrix with the spanning vectors of V as its columns: 01 11 A = @1 0A : 1 1 Then the plane V we're looking at is precisely the image of A, so the question is to find the vector 1 in this image, i.e. of the form A~x, closest to ~b = −1 . But this is precisely what the least-squares −2 method gives us! Fact. The solution ~x of AT A~x = AT~b, i.e. the so-called \least-squares solution" of A~x = ~b is the vector ~x so that A~x is as close as possible to ~b among all vectors in the image of A. In other words, the solution ~x is the one which makes A~x the orthogonal projection of ~b onto the image of A. 1 So, the vector ~x for which the vector A~x in the plane V = im A is closest to ~b = −1 is the −2 one which solves AT A~x = AT~b, which in this case is: 01 11 0 1 1 1 1 1 1 1 1 1 0 ~x = −1 : 1 0 1 @ A 1 0 1 @ A 1 1 −2 −1 1 Solving this gives ~x = , so the orthogonal projection of −1 onto the plane V = im A is 1=2 −2 01 11 0−1=21 −1 A~x = 1 0 = −1 ; @ A 1=2 @ A 1 1 −1=2 3 agreeing with what we found in the example doing it another way. The point is that while the least-squares method usually shows up in problems about best-fitting some given data points, it may also be useful in other types of length-minimizing problems; in particular ones dealing with subspaces which can be expressed as the image of some matrix. Fact. Many things dealing with minimizing some quantity or finding something which is closest to something else will involve either orthogonal projections or least squares. Orthogonal Matrices Now we move on to consider matrices analogous to the Q showing up in the formula for the matrix of an orthogonal projection. The difference now is that while Q from before was not necessarily a square matrix, here we consider ones which are square. Fact. The following are equivalent characterizations of an orthogonal matrix Q: • The columns of Q are orthonormal • QT = Q−1, which is the same as saying QT Q = I = QQT • Q is length-preserving or dot product preserving in the sense that computing lengths or dot products after multiplying vectors by Q gives the same result as computing these before mul- tiplying vectors by Q • Q describes geometrically either a rotation or a reflection and which one you get is determined by the sign of det Q. To elaborate on the last fact, note first that the second fact above implies that det Q = ±1 for any orthogonal matrix Q; then the rotations are the ones for which det Q = 1 and the reflections are the ones for which det Q = −1. Note that we are not saying that any matrix such that det A = 1 is a rotation or any one with det A = −1 is a reflection: this only applies to matrices we already know are orthogonal.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    7 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us