Numerical Optimization with CMA-ES: from Theory to Practice and from Practice to Theory

Numerical Optimization with CMA-ES: from Theory to Practice and from Practice to Theory

Numerical Optimization with CMA-ES: From Theory to Practice and from Practice to Theory Anne Auger Optimization and Machine Learning Team (TAO) INRIA Saclay-Ile-de-France Mascot-Num 8 - 10 April 2015 Note: some slides are taken from the CMA-ES tutorial given by A. Auger and N. Hansen at GECCO 2014 Zero Order Numerical / Continuous Black-Box Optimization Optimize f : Rn 7! R Zeroth order method + Black-Box setting x ∈ Rn f(x) ∈ R Gradients not available or not useful Knowledge about the problem encoded within the black box but not exploited by the algorithms Cost = # calls to the black-box (f-calls) Example of Numerical Black-box Problem Optimization of Oil Well Placement Z. Bouzarkouna PhD thesis, collab IFP ∈ R commercial solver NPV (in $) (source code not available) predict quantity of oil for given set of wells well configurations expensive optimization gradient not provided 3 Landscape of Numerical Zero Order Methods implicit assumption: # f-evals ≥ 100n Derivative-free optimization (deterministic) mathematical programming Trust-region methods (NEWUOA, BOBYQA) [Powell 2006, 2009] Simplex downhill [Nelder & Mead 1965] Pattern search [Hook and Jeeves 1961] Quasi-Newton with estimation of gradient (BFGS) [Broyden et al. 1970] Stochastic (randomized) search methods Evolutionary Algorithms (continuous domain) Differential Evolution [Storm & Price 1997] Particle Swarm Optimization [Kennedy & Eberhart 1995] Evolution Strategies [Rechenberg 1965, Hansen & Ostermeier 2001] Genetic Algorithms [Holland 1975, Goldberg 1989] Simulated annealing [Kirkpatrick et al. 1983] Simultaneous perturbation stochastic approximation (SPSA) [Spall 2000] 4 Landscape of Numerical Zero Order Methods Derivative-free optimization (deterministic) mathematical programming Trust-region methods (NEWUOA, BOBYQA) [Powell 2006, 2009] Simplex downhill [Nelder & Mead“local” 1965] optimization Pattern search [Hook and Jeeves 1961] Quasi-Newton with estimation of gradient (BFGS) [Broyden et al. 1970] Stochastic (randomized) search methods Evolutionary Algorithms (continuous domain) Differential Evolution [Storm & Price 1997] Particle Swarm Optimization“global” [Kennedy optimization & Eberhart 1995] Evolution Strategies [Rechenberg 1965, Hansen & Ostermeier 2001] Genetic Algorithms [Holland 1975, Goldberg 1989] Simulated annealing [Kirkpatrick et al. 1983] Simultaneous perturbation stochastic approximation (SPSA) [Spall 2000] 5 CMA-ES in a nutshell Covariance Matrix Adaptation Evolution Strategy N. Hansen main driving force behind Stochastic comparison-based algorithm Variable-metric method state-of-the-art stochastic Robust local search continuous optimizer Parameter-free algorithm Mature code available in C, Python, C++, Java, R, Matlab/Octave, Fortran - available at https://www.lri.fr/~hansen/cmaes_inmatlab.html algorithm: google Hansen + CMA-ES Overview General context What makes an optimization problem difficult? Insights into CMA-ES Adaptation of the mean vector Adaptation of the step-size Adaptation of the covariance matrix Variable metric illustration - learning inverse Hessian Local versus global search - comparisons with BFGS / NEWUOA Theoretical aspects Invariance Connexion with gradient optimization on manifolds (information geometry) Which Typical Difficulties Need to be Addressed? n x R x ∈ R ? f( ) ∈ The algorithm does not know in advance the features/ difficulties of the optimization problem that need to be solved difficulties related to real-wold problems Has to be ready to solve any of them What Makes a Function Difficult to S Why Comparison-based Stochastic S non-linear, non-quadratic, non convex ruggedness 100 90 non-smooth, discontinuous, multi-modal, and/or 80 70 60 noisy functions 50 40 30 20 10 0 dimensionality (size of the search space) −4 −3 −2 −1 0 1 2 3 4 3 (considerably) larger than three 2 curse of dimensionality 1 0 non-separability −1 −2 −3 dependencies between the objective variables −3 −2 −1 0 1 2 3 ill-conditioning Curse of Dimensionality Non-separability Non-separability (cont.) Ill-conditioned Problems Curvature of level sets Overview General context What makes an optimization problem difficult? Insights into CMA-ES Adaptation of the mean vector Adaptation of the step-size Adaptation of the covariance matrix Variable metric illustration - learning inverse Hessian Local versus global search - comparisons with BFGS / NEWUOA Theoretical aspects Invariance Connexion with gradient optimization on manifolds (information geometry) CMA-ES High-level Template ! : R → R ! θ!λ θ():,...,λ ∈ R ,...,λ θ ← (θ, ,...,λ, (),...,(λ)) Everything depends on p θ and F pθ proba. distribution encodes the belief where good solutions are located F should drive the algorithm to converge towards some optima of the objective function Sampling Distribution in Evolution Strategies Multivariate Normal Distributions m + σN (, C) − : θ m C ( )= exp ( m) C ( m)⇥ :=( , ) √(π)|C| − − − n m ∈ R favorite (incumbent) solution at a given iteration σ overall scaling - step-size of the algorithm m C Symmetric definite positive matrix - encodes the geometric shape of the distribution − Variable metric method: the multivariate {|( − m)C ( − m)=} normal distribution encodes the underlying metric of CMA-ES CMA-ES High-level Template ! : R → R ! θ!λ θ():,...,λ ∈ R ,...,λ θ ← (θ, ,...,λ, (),...,(λ)) Next: Insights into how m , σ , C are updated in CMA-ES Updating the mean vector m Invariance under Monotonically Increasing Transformations of f Comparison-based algorithm: Update of all parameters uses only ranking of solutions f(x1:λ) ≤ f(x2:λ) ≤ ...≤ f(xλ:λ) Same ranking on f or g ◦ f if g : R 7! R strictly increasing g ◦ f(x1:λ) g ◦ f(x2:λ) ... g ◦ f(xλ:λ) Invariance to strict. increasing transformations of f Overview General context What makes an optimization problem difficult? Insights into CMA-ES Adaptation of the mean vector Adaptation of the step-size Adaptation of the covariance matrix Variable metric illustration - learning inverse Hessian Local versus global search - comparisons with BFGS / NEWUOA Theoretical aspects Invariance Connexion with gradient optimization on manifolds (information geometry) Updating the Step-size Why step-size control? Overview General context What makes an optimization problem difficult? Insights into CMA-ES Adaptation of the mean vector Adaptation of the step-size Adaptation of the covariance matrix Variable metric illustration - learning inverse Hessian Local versus global search - comparisons with BFGS / NEWUOA Theoretical aspects Invariance Connexion with gradient optimization on manifolds (information geometry) Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) initial distribution, C = I ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) initial distribution, C = I ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) yw, movement of the population mean m (disregarding σ) ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) mixture of distribution C and step yw, T C ← 0.8 × C + 0.2 × ywyw ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) new distribution (disregarding σ) ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) new distribution (disregarding σ) ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) movement of the population mean m ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) mixture of distribution C and step yw, T C ← 0.8 × C + 0.2 × ywyw ...equations Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update µ m ← m + σyw, yw = Pi=1 wi yi:λ, yi ∼ Ni(0, C) new distribution, T C ← 0.8 × C + 0.2 × ywyw the ruling principle: the adaptation increases the likelihood of successful steps, yw, to appear again another viewpoint: the adaptation follows a natural gradient ...equations approximation of the expected fitness Anne Auger & Nikolaus Hansen CMA-ES July, 2014 45 / 81 Covariance Matrix Adaptation (CMA) Covariance Matrix Rank-One Update Covariance Matrix Adaptation Rank-One Update n 2 Initialize m ∈ R , and C = I, set σ = 1, learning rate ccov ≈ 2/n While not terminate xi = m + σ yi, yi ∼ Ni(0, C) , µ m ← m + σyw where yw = X wi yi:λ i=1 T 1 C ← 1 − ccov C ccov w ywy w ≥ 1 ( ) + µ w where µ = µ w 2 Pi=1 i rank-one|{z} The rank-one update has been found independently in several domains6 7 8 9 6 Kjellstrom&Tax¨ en´ 1981. Stochastic Optimization in System Design, IEEE TCS 7 Hansen&Ostermeier 1996. Adapting arbitrary normal mutation distributions in evolution strategies: The covariance matrix adaptation, ICEC 8 Ljung 1999. System Identification: Theory for the User 9 Haario et al 2001. An adaptive Metropolis algorithm, JSTOR Anne Auger & Nikolaus Hansen CMA-ES July, 2014 46 / 81 Covariance Matrix Adaptation (CMA) Cumulation—the Evolution Path Cumulation Utilizing the EvolutionT Path T T We used ywyw for updating C.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    66 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us