Barrier Method

Barrier Method

Barrier Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: Newton's method Consider the problem min f(x) x n for f convex, twice differentiable, with dom(f) = R . Newton's (0) n method: choose initial x R , repeat 2 (k) (k−1) 2 (k−1) −1 (k−1) x = x tk f(x ) f(x ); k = 1; 2; 3;::: − r r Step sizes tk chosen by backtracking line search If f Lipschitz, f strongly convex, 2f Lipschitz, then Newton's r r method has a local convergence rate O(log log(1/)) Downsides: Requires solving systems in Hessian quasi-Newton • Can only handle equality constraints this lecture • 2 Hierarchy of second-order methods Assuming all problems are convex, you can think of the following hierarchy that we've worked through: Quadratic problems are the easiest: closed-form solution • Equality-constrained quadratic problems are still easy: we use • KKT conditions to derive closed-form solution Equality-constrained smooth problems are next: use Newton's • method to reduce this to a sequence of equality-constrained quadratic problems Inequality- and equality-constrained smooth problems are • what we cover now: use interior-point methods to reduce this to a sequence of equality-constrained smooth problems 3 Log barrier function Consider the convex optimization problem min f(x) x subject to hi(x) 0; i = 1; : : : m ≤ Ax = b We will assume that f, h1; : : : hm are convex, twice differentiable, n each with domain R . The function m φ(x) = log( hi(x)) − − i=1 X is called the log barrier for the above problem. Its domain is the set of strictly feasible points, x : hi(x) < 0; i = 1; : : : m , which we assume is nonempty f g 4 Ignoring equality constraints for now, our problem can be written as m min f(x) + Ifh (x)≤0g(x) x i i=1 11.2 Logarithmic barrier function and central pathX 563 10 We approximate this representation by adding the log barrier function: 5 m min f(x) (1=t) log( hi(x)) 0 x − · − i=1 X 5 − 3 2 1 0 1 where t > 0 is a large number − − −u Figure 11.1 The dashed lines show the function I (u), and the solid curves − show I (u)= (1/t)log( u), for t =0.5, 1, 2. The curve for t =2gives − This approximation is more accurate for larger t. But for any value the best approximation.− − ! of t, the log barrier approaches if any hi(x) 0 1 ! The problem (11.3) has no inequality constraints, but its objective function is not (in general) differentiable, so Newton’s method cannot be applied. 5 11.2.1 Logarithmic barrier The basic idea of the barrier method is to approximate the indicator function I by the function − I (u)= (1/t)log( u), dom I = R++, − − − − − where t>0isaparameterthatsetstheaccuracyoftheapproximation.! ! Like I ,thefunctionI is convex and nondecreasing, and (by our convention) takes − − on the value for u>0. Unlike I ,however,I is differentiable and closed: it increases to∞ !as u increases to 0.− Figure 11.1− shows the function I ,and ∞ − the approximation I ,forseveralvaluesoft. As !t increases, the approximation becomes more accurate.− Substituting I !for I in (11.3) gives the approximation − − minimize f (x)+ m (1/t)log( f (x)) ! 0 i=1 i (11.4) subject to Ax = b. − − " The objective here is convex, since (1/t)log( u)isconvexandincreasinginu, and differentiable. Assuming an appropriate− closedness− condition holds, Newton’s method can be used to solve it. The function m φ(x)= log( f (x)), (11.5) − − i i=1 # Outline Today: Central path • Properties and interpretations • Barrier method • Convergence analysis • Feasibility methods • 6 Log barrier calculus For the log barrier function m φ(x) = log( hi(x)) − − i=1 X let us write down its gradient and Hessian, for future reference: m 1 φ(x) = hi(x) r − h (x)r i=1 i X and m m 2 1 T 1 2 φ(x) = hi(x) hi(x) hi(x) r h (x)2 r r − h (x)r i=1 i i=1 i X X computed using the chain rule 7 Central path Consider minimizing our problem, after replacing hard inequalities with barrier term: min tf(x) + φ(x) x subject to Ax = b (Here we switched placement of t, but its role is the same.) The central path is defined as the solution x?(t) as a function of t > 0. These solutions are characterized by the KKT conditions: ? ? Ax (t) = b; hi(x (t)) < 0; i = 1; : : : m m ? 1 ? T t f(x (t)) hi(x (t)) + A w = 0 r − h (x?(t))r i=1 i X m ? ? for some w R . As t , hope is that x (t) x , solution of our original2 problem ! 1 ! 8 As an important example, consider the barrier problem for a linear program: m T T min tc x log(ei d x) x − − i i=1 X The barrier function corresponds to polyhedral constraint Dx e ≤ 566 11 Interior-point methods Stationarity or centrality condition: m c 1 0 = tc T ? di − ei d x (t) i=1 − i X x⋆(10) x⋆ This means that gradient φ(x?(t)) r must be parallel to c, i.e., hyper- T T−? plane x : c x = c x (t) liesFigure tan- 11.2 Central path for an LP with n =2andm =6.Thedashed f ?g curves show(From three contour B lines & of theV logarithmic page 565) barrier function φ.The gent to contour of φ at x (t) central path converges to the optimal point x⋆ as t .Alsoshownisthe point on the central path with t =10.Theoptimalitycondition(11.9)at→∞ this point can be verified geometrically: The line cT x = cT x⋆(10) is tangent to the contour line of φ through x⋆(10). 9 we see that x⋆(t)minimizestheLagrangian m L(x, λ, ν)=f (x)+ λ f (x)+νT (Ax b), 0 i i − i=1 ! for λ = λ⋆(t)andν = ν⋆(t), which means that λ⋆(t), ν⋆(t)isadualfeasiblepair. Therefore the dual function g(λ⋆(t), ν⋆(t)) is finite, and m g(λ⋆(t), ν⋆(t)) = f (x⋆(t)) + λ⋆(t)f (x⋆(t)) + ν⋆(t)T (Ax⋆(t) b) 0 i i − i=1 ! = f (x⋆(t)) m/t. 0 − In particular, the duality gap associated with x⋆(t)andthedualfeasiblepairλ⋆(t), ν⋆(t)issimplym/t.Asanimportantconsequence,wehave f (x⋆(t)) p⋆ m/t, 0 − ≤ i.e., x⋆(t)isnomorethanm/t-suboptimal. This confirms the intuitive idea that x⋆(t)convergestoanoptimalpointast . →∞ Example 11.2 Inequality form linear programming. The dual of the inequality form LP (11.8) is maximize bT λ − subject to AT λ + c =0 λ 0. ≽ From the optimality conditions (11.9), it is clear that 1 λ⋆(t)= ,i=1,...,m, i T ⋆ t(bi a x (t)) − i Dual points from central path From central points, we can derive feasible dual points for original problem. Given x?(t) and corresponding w, we define ? 1 ? ui (t) = ? ; i = 1; : : : m; v (t) = w=t −thi(x (t)) We claim u?(t); v?(t) are dual feasible for original problem. Why? Note that u?(t) > 0 since h (x?(t)) < 0 for all i • i i Further, the point (u?(t); v?(t)) lies in domain of Lagrange • dual function g(u; v), since by definition m ? ? ? T ? f(x (t)) + ui(x (t)) hi(x (t)) + A v (t) = 0 r r i=1 X I.e., x?(t) minimizes Lagrangian L(x; u?(t); v?(t)) over x, so g(u?(t); v?(t)) > −∞ 10 This allows us to bound suboptimality of f(x?(t)), with respect to original problem, via the duality gap. We compute m ? ? ? ? ? g(u (t); v (t)) = f(x (t)) + ui (t)hi(x (t)) + i=1 X v?(t)T (Ax?(t) b) − = f(x?(t)) m=t − That is, we know that f(x?(t)) f ? m=t − ≤ This will be very useful as a stopping criterion; it also confirms the fact that x?(t) x? as t ! ! 1 11 Interpretation via perturbed KKT conditions We can think of central path solution x?(t) and corresponding dual point (u?(t); v?(t)) as solving the perturbed KKT conditions m ? ? ? T ? f(x (t)) + ui(x (t)) hi(x (t)) + A v (t) = 0 r r i=1 ? X ? ui (t) hi(x (t)) = 1=t; i = 1; : : : m ? · − ? hi(x (t)) 0; i = 1; : : : m; Ax (t) = b ≤? ui(x (t)) 0; i = 1; : : : m ≥ Only difference between these and actual KKT conditions for our original problem is in the second condition: these are replaced by ? ? u (t) hi(x (t)) = 0; i = 1; : : : m i · i.e., complementary slackness, in actual KKT conditions 12 First attempt at an algorithm Since we have seen that solution x?(t) of min tf(x) + φ(x) x subject to Ax = b is no more than m=t suboptimal, why don't we simply pick desired accuracy level , set t = m/, and then solve above problem using Newton's method? This is like directly seeking out a point near the end of the central path. Problem is that the required t is often huge, and this causes serious numerical issues in practice. Hence this approach is almost never used A better approach is to traverse the entire central path, in order to reach the end 13 Caveat: what does this remind you of? The central path is closely related to the solution path of statistical optimization problems, defined over a tuning parameter t=1.22 ● 566 11 Interior-point methods t=0.91 ● c t=0.52 ● t=0.17 ● ● x⋆(10) x⋆ LP central path Ridge regression solution path Figure 11.2 Central path for an LP with n =2andm =6.Thedashed curves show three contour lines of the logarithmic barrier function φ.The Solving a statistical problem over acentral grid path of converges tuning to the optimal parameter point x⋆ as t values.Alsoshownisthe point on the central path with t =10.Theoptimalitycondition(11.9)at→∞ with warm starts is deeply connectedthis point to can the be verified central geometrically: path The concept line cT x = cT x⋆(10) is tangent ⋆ (but this connection is not as developedto the contour as line it of perhapsφ through x (10).

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    23 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us