Lecture 8 Least-Norm Solutions of Undetermined Equations

Lecture 8 Least-Norm Solutions of Undetermined Equations

EE263 Autumn 2007-08 Stephen Boyd Lecture 8 Least-norm solutions of undetermined equations • least-norm solution of underdetermined equations • minimum norm solutions via QR factorization • derivation via Lagrange multipliers • relation to regularized least-squares • general norm minimization with equality constraints 8–1 Underdetermined linear equations we consider y = Ax where A ∈ Rm×n is fat (m < n), i.e., • there are more variables than equations • x is underspecified, i.e., many choices of x lead to the same y we’ll assume that A is full rank (m), so for each y ∈ Rm, there is a solution set of all solutions has form { x | Ax = y } = { xp + z | z ∈ N (A) } where xp is any (‘particular’) solution, i.e., Axp = y Least-norm solutions of undetermined equations 8–2 • z characterizes available choices in solution • solution has dim N (A)= n − m ‘degrees of freedom’ • can choose z to satisfy other specs or optimize among solutions Least-norm solutions of undetermined equations 8–3 Least-norm solution one particular solution is T T −1 xln = A (AA ) y (AAT is invertible since A full rank) in fact, xln is the solution of y = Ax that minimizes kxk i.e., xln is solution of optimization problem minimize kxk subject to Ax = y (with variable x ∈ Rn) Least-norm solutions of undetermined equations 8–4 suppose Ax = y, so A(x − xln) = 0 and T T T T −1 (x − xln) xln = (x − xln) A (AA ) y T T −1 = (A(x − xln)) (AA ) y = 0 i.e., (x − xln) ⊥ xln, so 2 2 2 2 2 kxk = kxln + x − xlnk = kxlnk + kx − xlnk ≥ kxlnk i.e., xln has smallest norm of any solution Least-norm solutions of undetermined equations 8–5 { x | Ax = y } N (A)= { x | Ax = 0 } xln • orthogonality condition: xln ⊥ N (A) • projection interpretation: xln is projection of 0 on solution set { x | Ax = y } Least-norm solutions of undetermined equations 8–6 • A† = AT (AAT )−1 is called the pseudo-inverse of full rank, fat A • AT (AAT )−1 is a right inverse of A • I − AT (AAT )−1A gives projection onto N (A) cf. analogous formulas for full rank, skinny matrix A: • A† =(AT A)−1AT • (AT A)−1AT is a left inverse of A • A(AT A)−1AT gives projection onto R(A) Least-norm solutions of undetermined equations 8–7 Least-norm solution via QR factorization find QR factorization of AT , i.e., AT = QR, with n×m T • Q ∈ R , Q Q = Im • R ∈ Rm×m upper triangular, nonsingular then T T −1 −T • xln = A (AA ) y = QR y −T • kxlnk = kR yk Least-norm solutions of undetermined equations 8–8 Derivation via Lagrange multipliers • least-norm solution solves optimization problem minimize xT x subject to Ax = y • introduce Lagrange multipliers: L(x, λ)= xT x + λT (Ax − y) • optimality conditions are T ∇xL = 2x + A λ = 0, ∇λL = Ax − y = 0 • from first condition, x = −AT λ/2 • substitute into second to get λ = −2(AAT )−1y • hence x = AT (AAT )−1y Least-norm solutions of undetermined equations 8–9 Example: transferring mass unit distance f • unit mass at rest subject to forces xi for i − 1 < t ≤ i, i = 1,..., 10 • y1 is position at t = 10, y2 is velocity at t = 10 2 10 • y = Ax where A ∈ R × (A is fat) • find least norm force that transfers mass unit distance with zero final velocity, i.e., y = (1, 0) Least-norm solutions of undetermined equations 8–10 0.06 0.04 0.02 ln x 0 −0.02 −0.04 −0.06 0 2 4 6 8 10 12 t 1 0.8 0.6 0.4 position 0.2 0 0 2 4 6 8 10 12 t 0.2 0.15 0.1 0.05 velocity 0 0 2 4 6 8 10 12 t Least-norm solutions of undetermined equations 8–11 Relation to regularized least-squares • suppose A ∈ Rm×n is fat, full rank 2 2 • define J1 = kAx − yk , J2 = kxk • least-norm solution minimizes J2 with J1 = 0 2 2 • minimizer of weighted-sum objective J1 + µJ2 = kAx − yk + µkxk is 1 T − T xµ = A A + µI A y • fact: xµ → xln as µ → 0, i.e., regularized solution converges to least-norm solution as µ → 0 • in matrix terms: as µ → 0, −1 −1 AT A + µI AT → AT AAT (for full rank, fat A) Least-norm solutions of undetermined equations 8–12 General norm minimization with equality constraints consider problem minimize kAx − bk subject to Cx = d with variable x • includes least-squares and least-norm problems as special cases • equivalent to minimize (1/2)kAx − bk2 subject to Cx = d • Lagrangian is L(x, λ) = (1/2)kAx − bk2 + λT (Cx − d) = (1/2)xT AT Ax − bT Ax + (1/2)bT b + λT Cx − λT d Least-norm solutions of undetermined equations 8–13 • optimality conditions are T T T ∇xL = A Ax − A b + C λ = 0, ∇λL = Cx − d = 0 • write in block matrix form as AT A CT x AT b = C 0 λ d • if the block matrix is invertible, we have −1 x AT A CT AT b = λ C 0 d Least-norm solutions of undetermined equations 8–14 if AT A is invertible, we can derive a more explicit (and complicated) formula for x • from first block equation we get x =(AT A)−1(AT b − CT λ) • substitute into Cx = d to get C(AT A)−1(AT b − CT λ)= d so −1 λ = C(AT A)−1CT C(AT A)−1AT b − d • recover x from equation above (not pretty) −1 x =(AT A)−1 AT b − CT C(AT A)−1CT C(AT A)−1AT b − d Least-norm solutions of undetermined equations 8–15.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    15 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us