Convex Function with Lipschitz Continuous Gradient

Convex Function with Lipschitz Continuous Gradient

Lecture 2 Gradient Descent and Subgradient Methods Jie Lu ([email protected]) Richard Combes Alexandre Proutiere Automatic Control, KTH September 12, 2013 Outline Convex analysis Gradient descent method Gradient projection method Subgradient method 2 Convex Set Examples . Hyperplane . Polyhedral . Ellipsoid Convex hull: the set of all convex combinations of the points in X . Convex Combination: 3 Convex Function . 4 Strong Convexity 5 Convex Function with Lipschitz Continuous Gradient f is convex and has Lipschitz continuous gradient iff one of the following holds: If f is strongly convex and has Lipschitz continuous gradient, then 6 Gradient Descent Method Unconstrained convex optimization . f is convex and continuously differentiable Optimality condition: . Unique if f is strictly convex Basic gradient method A descent method (for sufficiently small stepsize ) 7 Convergence Analysis Choose sufficiently small stepsize so that . Need further assumptions to guarantee convergence . For constant stepsize . For diminishing stepsize . 8 Convergence Rate Suppose f has Lipschitz continuous gradient with L > 0 and use constant stepsize Then, . R.H.S. achieves minimum when Further suppose f is strongly convex with and use constant stepsize . Then, . q achieves minimum when . 9 Gradient Projection Method Constrained convex optimization . f is convex and continuously differentiable . X is a nonempty, closed, and convex set Optimality condition . Unique if f is strictly convex Gradient projection method . Projection operator . Similar convergence analysis as unconstrained case, using properties of projection . 10 Important Facts of Projection 11 Subgradient and Subdifferential Consider a convex and possibly non-differentiable function . . 12 Subgradient Method Consider the (constrained) nonsmooth convex optimization problem Optimality condition . Subgradient method 13 Convergence Analysis . Very similar to the convergence analysis of the gradient descent method Averages behave better 14 Summary Convex set Convex function . Strictly convex, strongly convex, Lipschitz continuous gradient Gradient descent method . Smooth unconstrained convex optimization . Convergence performance . Lipschitz continuous function: . Lipschitz continuous gradient: sublinear convergence . Strongly convex function with Lipschitz continuous gradient: linear convergence Gradient projection method . Smooth constrained convex optimization . Facts of projection . Similar convergence results as gradient descent method Subgradient method . Subgradient and subdifferential . Nonsmooth convex optimization . Convergence complexity 15 References Y. Nesterov, Introductory lectures on Convex Optimization: A Basic Course. Norwell, MA: Kluwer Academic Publishers, 2004. D. P. Bertsekas, Nonlinear Programming. Belmont, MA: Athena Scientific, 1999. D. P. Bertsekas, A. Nedich, and A. Ozdaglar, Convex Analysis and Optimization. Belmont, MA: Athena Scientific, 2003. S. Boyd and L. Vandenberghe, Convex Optimization. New York, NY: Cambridge University Press, 2004. 16 .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    16 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us