11. Quadratic Forms and Ellipsoids

11. Quadratic Forms and Ellipsoids

CS/ECE/ISyE 524 Introduction to Optimization Spring 2017{18 11. Quadratic forms and ellipsoids Quadratic forms Orthogonal decomposition Positive definite matrices Ellipsoids Laurent Lessard (www.laurentlessard.com) Quadratic forms Linear functions: sum of terms of the form ci xi where the ci are parameters and xi are variables. General form: T c1x1 + ··· + cnxn = c x Quadratic functions: sum of terms of the form qij xi xj where qij are parameters and xi are variables. General form: 2 2 2 q11x1 + q12x1x2 + ··· + qnnxn (n terms) 2 3T 2 3 2 3 x1 q11 ::: q1n x1 6 . 7 6 . .. 7 6 . 7 T = 4 . 5 4 . 5 4 . 5 = x Qx xn qn1 ::: qnn xn 11-2 Quadratic forms Example: 4x 2 + 6xy − 2yz + y 2 − z 2 2x3T 2460 3 2x3 4y5 4010 5 4y5 z 0 −2 −1 z 2 3T 2 3 2 3 8 x 4 p2 q2 x < p1 + p2 = 6 In general: 4y5 4p1 1 r2 5 4y5 q1 + q2 = 0 z q1 r1 −1 z : r1 + r2 = −2 2x3T 24 3 0 3 2x3 Symmetric: 4y5 43 1 −15 4y5 z 0 −1 −1 z 11-3 Quadratic forms Any quadratic function f (x1;:::; xn) can be written in the form x TQx where Q is a symmetric matrix (Q = QT). T Proof: Suppose f (x1;:::; xn) = x Rx where R is not symmetric. Since it is a scalar, we can take the transpose: T x TRx = x TRx = x TRTx Therefore: T 1 T T T T 1 T x Rx = 2 x Rx + x R x = x 2 (R + R )x 1 T So we're done, because 2 (R + R ) is symmetric! 11-4 Orthogonal decomposition Theorem. Every real symmetric matrix Q = QT 2 Rn×n can be decomposed into a product: Q = UΛUT where Λ = diag(λ1; : : : ; λn) is a real diagonal matrix, and U 2 Rn×n is an orthogonal matrix. i.e. it satisfies UTU = I . This is a useful decomposition because orthogonal matrices have very nice properties... 11-5 Orthogonal matrices A matrix U is orthogonal if UTU = I . If the columns are U = u1 u2 ··· um , then we have: 2 T T 3 2 3 u1 u1 ··· u1 um 1 ··· 0 T 6 . .. 7 6. .. .7 U U = 4 . 5 = 4. .5 T T umu1 ··· umum 0 ··· 1 T Columns of U are mutually orthogonal: ui uj = 0 if i 6= j. If U is square, U−1 = UT, and UT is also orthogonal. 11-6 Orthogonal matrices columns can be rearranged and the factorization stays valid. 2 3 2 T3 λ1 0 0 u1 T u1 u2 u3 4 0 λ2 0 5 4u2 5 T 0 0 λ3 u3 T T T = λ1u1u1 + λ2u2u2 + λ3u3u3 2 3 2 T3 λ1 0 0 u1 T = u1 u3 u2 4 0 λ3 0 5 4u3 5 T 0 0 λ2 u2 11-7 Orthogonal matrices Orthogonal matrices preserve angle and (2-norm) distance: (Ux)T(Uy) = x T(UTU)y = x Ty In particular, we have kUzk = kzk for any z. T If Q = UΛU , then multiply by ui : 2 T3T 2 3 2 T3 u1 λ1 ··· 0 u1 6 . 7 6 . .. 7 6 . 7 Qui = 4 . 5 4 . 5 4 . 5 ui = λi ui T T un 0 ··· λn un So multiplication by Q simply scales each ui by λi . In other words: (λi ; ui ) are the eigenvalue-eigenvector pairs of Q. 11-8 Orthogonal matrix example Rotation matrices are orthgonal: cos θ sin θ R = θ − sin θ cos θ We can verify this: cos θ − sin θ cos θ sin θ RTR = θ θ sin θ cos θ − sin θ cos θ cos2 θ + sin2 θ cos θ sin θ − sin θ cos θ = sin θ cos θ − cos θ sin θ sin2 θ + cos2 θ 1 0 = 0 1 T Note: Rθ = R−θ. This holds for 3D rotation matrices also... 11-9 Eigenvalues and eigenvectors If A 2 Rn×n and there is a vector v and scalar λ such that Av = λv Then v is an eigenvector of A and λ is the corresponding eigenvalue. Some facts: Any square matrix has n eigenvalues. Each eigenvalue has at least one corresponding eigenvector. In general, eigenvalues & eigenvectors can be complex. In general, eigenvectors aren't orthogonal, and may not even be linearly independent. i.e. V = v1 ··· vn may not be invertible. If it is, we say that A is diagonalizable and then A = V ΛV −1. Otherwise, Jordan Canonical Form. Symmetric matrices are much simpler! 11-10 Recap: symmetric matrices T n×n Every symmetric Q = Q 2 R has n real eigenvalues λi . There exist n mutually orthogonal eigenvectors u1;:::; un: Qui = λi ui for all i = 1;:::; n ( T 1 if i = j u uj = i 0 if i 6= j T If we define U = u1 ··· un then U U = I and 2 3 λ1 ··· 0 6 . .. 7 T Q = U 4 . 5 U 0 ··· λn 11-11 Eigenvalue example Consider the quadratic: 7x 2 + 4xy + 6y 2 + 4yz + 5z 2. A simple question: are there values that make this negative? 2x3T 27 2 03 2x3 equivalent to: 4y5 42 6 25 4y5 z 0 2 5 z Orthogonal decomposition: 2 3 2 1 2 2 3 2 3 2 1 2 2 3T 7 2 0 − 3 3 3 3 0 0 − 3 3 3 2 6 2 = 6 2 − 1 2 7 60 6 07 6 2 − 1 2 7 4 5 4 3 3 3 5 4 5 4 3 3 3 5 0 2 5 2 2 1 2 2 1 − 3 − 3 3 0 0 9 − 3 − 3 3 Eigenvalues are f3; 6; 9g. 11-12 Eigenvalue example Eigenvalue decomposition: 2 3 2 1 2 2 3 2 3 2 1 2 2 3T 7 2 0 − 3 3 3 3 0 0 − 3 3 3 2 6 2 = 6 2 − 1 2 7 60 6 07 6 2 − 1 2 7 4 5 4 3 3 3 5 4 5 4 3 3 3 5 0 2 5 2 2 1 2 2 1 − 3 − 3 3 0 0 9 − 3 − 3 3 Define new coordinates: 2 3 2 1 2 2 3T 2 3 p − 3 3 3 x q = 6 2 − 1 2 7 y 4 5 4 3 3 3 5 4 5 r 2 2 1 z − 3 − 3 3 Then we can write: 2x3T 27 2 03 2x3 2p3T 23 0 03 2p3 4y5 42 6 25 4y5 = 4q5 40 6 05 4q5 z 0 2 5 z r 0 0 9 r 11-13 Eigenvalue example After some manipulations, we discovered that 7x 2 + 4xy + 6y 2 + 4yz + 5z 2 = 3p2 + 6q2 + 9r 2 where: 1 2 2 p = − 3 x + 3 y − 3 z 2 1 2 q = 3 x − 3 y − 3 z 2 2 1 r = 3 x + 3 y + 3 z Conclusion: the quadratic can never be negative. 11-14 Recap Question: Is x TQx ever negative? Answer: Look at the orthogonal decomposition of Q: Q = UΛUT Define new coordinates z = UTx. T 2 2 x Qx = λ1z1 + ··· + λnzn T If all λi ≥ 0, then x Qx ≥ 0 for any x. If some λk < 0, set zk = 1 and all other zi = 0. Then find corresponding x using x = Uz, and x TQx < 0. 11-15 Positive definite matrices For a matrix Q = QT, the following are equivalent: 1. x TQx ≥ 0 for all x 2 Rn 2. all eigenvalues of Q satisfy λi ≥ 0 A matrix with this property is called positive semidefinite (PSD). The notation is Q 0. Note: When we talk about PSD matrices, we always assume we're talking about a symmetric matrix. 11-16 Positive definite matrices Name Definition Notation Positive semidefinite all λi ≥ 0 Q 0 Positive definite all λi > 0 Q 0 Negative semidefinite all λi ≤ 0 Q 0 Negative definite all λi < 0 Q ≺ 0 Indefinite everything else (none) Some properties: If P 0 then −P 0 If P 0 and α > 0 then αP 0 If P 0 and Q 0 then P + Q 0 Every R = RT can be written as R = P − Q for some appropriate choice of matrices P 0 and Q 0. 11-17 Ellipsoids For linear constraints, the set of x satisfying cTx = b is a hyperplane and the set cTx ≤ b is a halfspace. For quadratic constraints: If Q 0, the set x TQx ≤ b is an ellipsoid. 11-18 Ellipsoids By orthogonal decomposition, we can write x TQx = z TΛz where we defined the new coordinates z = UTx. The set of x satisfying x TQx ≤ 1 corresponds to the set of 2 2 z satisfying λ1z1 + ··· + λnzn ≤ 1. If Q 0, then λi > 0. In the z coordinates, this is a stretched sphere (ellipsoid). In the zi direction, it is stretched by p1 . λi Since x = Uz, and this transformation preserves angles and distances (think of it as a rotation), then in the xi coordinates, it is a rotated ellipsoid. The principal axes (the zi directions) map to the ui directions after the rotation. 11-19 Ellipsoids Plot of the region 3p2 + 6q2 + 9r 2 ≤ 1 Ellipse axes are in the directions e1, e2, e3 Plot of the region 7x 2 + 4xy + 6y 2 + 4yz + 5z 2 ≤ 1 Ellipse axes are in the directions u1, u2, u3 11-20 Norm representation If Q 0 we can define the matrix square root: 1. Let Q = UΛUT be an orthogonal decomposition p p 1=2 2.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    21 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us