Numericaloptimization

Total Page:16

File Type:pdf, Size:1020Kb

Numericaloptimization Numerical Optimization Alberto Bemporad http://cse.lab.imtlucca.it/~bemporad/teaching/numopt Academic year 2020-2021 Course objectives Solve complex decision problems by using numerical optimization Application domains: • Finance, management science, economics (portfolio optimization, business analytics, investment plans, resource allocation, logistics, ...) • Engineering (engineering design, process optimization, embedded control, ...) • Artificial intelligence (machine learning, data science, autonomous driving, ...) • Myriads of other applications (transportation, smart grids, water networks, sports scheduling, health-care, oil & gas, space, ...) ©2021 A. Bemporad - Numerical Optimization 2/102 Course objectives What this course is about: • How to formulate a decision problem as a numerical optimization problem? (modeling) • Which numerical algorithm is most appropriate to solve the problem? (algorithms) • What’s the theory behind the algorithm? (theory) ©2021 A. Bemporad - Numerical Optimization 3/102 Course contents • Optimization modeling – Linear models – Convex models • Optimization theory – Optimality conditions, sensitivity analysis – Duality • Optimization algorithms – Basics of numerical linear algebra – Convex programming – Nonlinear programming ©2021 A. Bemporad - Numerical Optimization 4/102 References i ©2021 A. Bemporad - Numerical Optimization 5/102 Other references • Stephen Boyd’s “Convex Optimization” courses at Stanford: http://ee364a.stanford.edu http://ee364b.stanford.edu • Lieven Vandenberghe’s courses at UCLA: http://www.seas.ucla.edu/~vandenbe/ • For more tutorials/books see http://plato.asu.edu/sub/tutorials.html ©2021 A. Bemporad - Numerical Optimization 6/102 Optimization modeling What is optimization? • Optimization = assign values to a set of decision variables so to optimize a certain objective function • Example: Which is the best velocity to minimize fuel consumption ? fuel [ℓ/km] velocity [km/h] 0 30 60 90 120 160 ©2021 A. Bemporad - Numerical Optimization 7/102 What is optimization? • Optimization = assign values to a set of decision variables so to optimize a certain objective function • Example: Which is the best velocity to minimize fuel consumption ? best fuel fuel consumption [ℓ/km] optimal velocity velocity [km/h] 0 30 60 90 120 160 optimization variable: velocity cost function to minimize: fuel consumption parameters of the decision problem: engine type, chassis shape, gear, … ©2021 A. Bemporad - Numerical Optimization 8/102 Optimization problem f(x) min f(x) x f(x*) x ∗ x* f = minx f(x) = optimal value ∗ n n x = arg minx f(x) = optimizer x 2 R ; f : R ! R 2 3 x 6 1 7 maxx f(x) 6 . 7 x = 4 . 5 ; f(x) = f(x1; x2; : : : ; xn) xn Most often the problem is difficult to solve by inspection use a numerical solver implementing an optimization algorithm ©2021 A. Bemporad - Numerical Optimization 9/102 Optimization problem min f(x) x • The objective function f : Rn ! R models our goal: minimize (or maximize) some quantity. For example fuel, money, distance from a target, etc. • The optimization vector x 2 Rn is the vector of optimization variables (or unknowns) xi to be decided optimally. For example velocity, number of assets in a portfolio, voltage applied to a motor, etc. ©2021 A. Bemporad - Numerical Optimization 10/102 Constrained optimization problem • The optimization vector x may not be completely free, but rather restricted to a feasible set X ⊆ Rn • Example: the velocity must be smaller than 60 km/h fuel [ℓ/km] velocity [km/h] 0 30 60 90 120 160 best fuel fuel consumption [ℓ/km] optimal velocity velocity [km/h] 0 30 60 90 120 160 The new optimizer is x∗ = 42 km/h. ©2021 A. Bemporad - Numerical Optimization 11/102 Constrained optimization problem f(x) minx f(x) s:t: g(x) ≤ 0 h(x) = 0 g(x) 0 x • The (in)equalities define the feasible set X of admissible variables g : Rn ! Rm; h : Rn ! Rp " # g (x ;x ;:::;x ) X = fx 2 Rn : g(x) ≤ 0; h(x) = 0g 1 1 2 n g(x) = . • Further constraints may restrict X , gm(x1;x2;:::;xn) " # for example: h1(x1;x2;:::;xn) n h(x) = . x 2 f0; 1g (x = binary vector) . x 2 Zn (x = integer vector) hp(x1;x2;:::;xn) ©2021 A. Bemporad - Numerical Optimization 12/102 A few observations • An optimization problem can be always written as a minimization problem max f(x) = − min{−f(x)g x2X x2X • Similarly, an inequality gi(x) ≥ 0 is equivalent to −gi(x) ≤ 0 • An equality h(x) = 0 is equivalent to the double inequalities h(x) ≤ 0, −h(x) ≤ 0 (often this is only good in theory, but not numerically) • Scaling f(x) to αf(x) and/or gi(x) to βigi(x), or shifting to f(x) + γ, does not change the optimizer, for all α; βi > 0 and γ. Same if hj(x) is scaled to γjhj(x), 8γj =6 0 • Adding constraints makes the objective worse or equal: min f(x) ≤ min f(x) x2X1 x2X1; x2X2 • Strict inequalities gi(x) < 0 can be approximated by gi(x) ≤ −ϵ (0 < ϵ ≪ 1) ©2021 A. Bemporad - Numerical Optimization 13/102 Infeasibility and unboundedness • A vector x 2 Rn is feasible if x 2 X , i.e., it satisfies the given constraints • A problem is infeasible if X = ; (the constraints are too tight) • A problem is unbounded if 8M > 0 9x 2 X such that f(x) < −M. In this case we write inf f(x) = −∞ x2X ©2021 A. Bemporad - Numerical Optimization 14/102 Global and local minima • A vector x∗ 2 Rn is a global optimizer if x 2 X and f(x) ≥ f(x∗), 8x 2 X • A vector x∗ 2 Rn is a strict global optimizer if x∗ 2 X and f(x) > f(x∗), 8x 2 X , x =6 x∗ • A vector x∗ 2 Rn is a (strict) local optimizer if x∗ 2 X and there exists a neighborhood1 N of x∗ such that f(x) ≥ f(x∗), 8x 2 X \ N (f(x) > f(x∗), 8x 2 X \ N , x =6 x∗) 1Neighborhood of x = open set containing x ©2021 A. Bemporad - Numerical Optimization 15/102 Example: Least Squares • We have a dataset (uk; yk), uk; yk 2 R, k = 1;:::N • We want to fit a line y^ = au + b to the dataset that minimizes 2 3 " y # 2 XN XN u1 1 1 0 . 2 uk 2 4 . 5 . f(x) = (yk − auk − b) = ([ ] x − yk) = . x − . 1 . k=1 k=1 uN 1 yN 2 a with respect to x = [ b ] a∗ a ∗ ∗ • The problem b∗ = arg min f([ b ]) is a least-squares problem: y^ = a u + b In MATLAB: 1.5 1 x=[u ones(size(u))]ny 0.5 0 y -0.5 In Python: -1 import numpy as np -1.5 -1 -0.5 0 0.5 1 A=np.hstack((u,np.ones(u.shape))) u x=np.linalg.lstsq(A,y,rcond=0)[0] ©2021 A. Bemporad - Numerical Optimization 16/102 Least Squares using Basis Functions • More generally: we can fit nonlinear functions y = f(u) expressed as the sum Xn of basis functions yk ≈ xiϕi(uk) using least squares i=1 2 3 4 • Example: fit polynomial function y = x1 + x2u1 + x3u1 + x4u1 + x5u1 XN h i 2 2 3 4 min yk − 1 uk u u u x least squares x k k k k=1 | {z } linear with respect to x 2 3 1 2.5 6 7 6 u1 7 6 7 2 6 2 7 ϕ(u) = 6 u1 7 4 u3 5 1 1.5 4 u1 1 0 0.5 1 1.5 2 ©2021 A. Bemporad - Numerical Optimization 17/102 Least Squares - Fitting a circle • Example: fit a circle to a set of data2 XN 2 2 2 2 min (r − (xk − x0) − (yk − y0) ) x0;y0;r k=1 x0 • Let x = y0 be the optimization vector (note the change of variables!) 2− 2− 2 r x0 y0 • The problem becomes the least squares problem 2 N h i X 2 1 − 2 2 min 2xk 2yk 1 x (xk + yk) x 0 k=1 -1 -2 -3 -2 -1 0 1 2 3 2http://www.utc.fr/~mottelet/mt94/leastSquares.pdf ©2021 A. Bemporad - Numerical Optimization 18/102 Convex sets Definition n A set S ⊆ R is convex if for all x1; x2 2 S λx1 + (1 − λ)x2 2 S; 8λ 2 [0; 1] convex set nonconvex set S x1 x2 x1 x2 ©2021 A. Bemporad - Numerical Optimization 19/102 Convex functions • f : S ! R is a convex function if S is convex and f(λx1 + (1 − λ)x2) ≤ λf(x1) + (1 − λ)f(x2) 8x1; x2 2 S; λ 2 [0; 1] Jensen’s inequality (Jensen, 1906) 3 • If f is convex and differentiable at x2, take the limit λ ! 0 and get 0 f(x1) ≥ f(x2) + rf(x2) (x1 − x2) Johan Jensen (1859–1925) • A function f is strictly convex if f(λx1 + (1 − λ)x2) < λf(x1) + (1 − λ)f(x2), 8x1 =6 x2 2 S, 8λ 2 (0; 1) 3 0 f(x1) − f(x2) ≥ limλ!0(f(x2 + λ(x1 − x2)) − f(x2))/λ = rf (x2)(x1 − x2) ©2021 A. Bemporad - Numerical Optimization 20/102 Convex functions • A function f : S ! R is strongly convex with parameter m ≥ 0 if mλ(1 − λ) f(λx + (1 − λ)x ) ≤ λf(x ) + (1 − λ)f(x ) − kx − x k2 1 2 1 2 2 1 2 2 • If f strongly convex with parameter m ≥ 0 and differentiable then m f(y) ≥ f(x) + rf(x)0(y − x) + ky − xk2 2 2 • Equivalently, f is strongly convex with parameter m ≥ 0 if and only if − m 0 f(x) 2 x x convex • Moreover, if f is differentiable twice this is equivalent to r2f(x) ≽ mI (i.e., matrix r2f(x) − mI is positive semidefinite), 8x 2 Rn • A function f is (strictly/strongly) concave if −f is (strictly/strongly) convex ©2021 A.
Recommended publications
  • University of California, San Diego
    UNIVERSITY OF CALIFORNIA, SAN DIEGO Computational Methods for Parameter Estimation in Nonlinear Models A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy in Physics with a Specialization in Computational Physics by Bryan Andrew Toth Committee in charge: Professor Henry D. I. Abarbanel, Chair Professor Philip Gill Professor Julius Kuti Professor Gabriel Silva Professor Frank Wuerthwein 2011 Copyright Bryan Andrew Toth, 2011 All rights reserved. The dissertation of Bryan Andrew Toth is approved, and it is acceptable in quality and form for publication on microfilm and electronically: Chair University of California, San Diego 2011 iii DEDICATION To my grandparents, August and Virginia Toth and Willem and Jane Keur, who helped put me on a lifelong path of learning. iv EPIGRAPH An Expert: One who knows more and more about less and less, until eventually he knows everything about nothing. |Source Unknown v TABLE OF CONTENTS Signature Page . iii Dedication . iv Epigraph . v Table of Contents . vi List of Figures . ix List of Tables . x Acknowledgements . xi Vita and Publications . xii Abstract of the Dissertation . xiii Chapter 1 Introduction . 1 1.1 Dynamical Systems . 1 1.1.1 Linear and Nonlinear Dynamics . 2 1.1.2 Chaos . 4 1.1.3 Synchronization . 6 1.2 Parameter Estimation . 8 1.2.1 Kalman Filters . 8 1.2.2 Variational Methods . 9 1.2.3 Parameter Estimation in Nonlinear Systems . 9 1.3 Dissertation Preview . 10 Chapter 2 Dynamical State and Parameter Estimation . 11 2.1 Introduction . 11 2.2 DSPE Overview . 11 2.3 Formulation . 12 2.3.1 Least Squares Minimization .
    [Show full text]
  • An Algorithm Based on Semidefinite Programming for Finding Minimax
    Takustr. 7 Zuse Institute Berlin 14195 Berlin Germany BELMIRO P.M. DUARTE,GUILLAUME SAGNOL,WENG KEE WONG An algorithm based on Semidefinite Programming for finding minimax optimal designs ZIB Report 18-01 (December 2017) Zuse Institute Berlin Takustr. 7 14195 Berlin Germany Telephone: +49 30-84185-0 Telefax: +49 30-84185-125 E-mail: [email protected] URL: http://www.zib.de ZIB-Report (Print) ISSN 1438-0064 ZIB-Report (Internet) ISSN 2192-7782 An algorithm based on Semidefinite Programming for finding minimax optimal designs Belmiro P.M. Duarte a,b, Guillaume Sagnolc, Weng Kee Wongd aPolytechnic Institute of Coimbra, ISEC, Department of Chemical and Biological Engineering, Portugal. bCIEPQPF, Department of Chemical Engineering, University of Coimbra, Portugal. cTechnische Universität Berlin, Institut für Mathematik, Germany. dDepartment of Biostatistics, Fielding School of Public Health, UCLA, U.S.A. Abstract An algorithm based on a delayed constraint generation method for solving semi- infinite programs for constructing minimax optimal designs for nonlinear models is proposed. The outer optimization level of the minimax optimization problem is solved using a semidefinite programming based approach that requires the de- sign space be discretized. A nonlinear programming solver is then used to solve the inner program to determine the combination of the parameters that yields the worst-case value of the design criterion. The proposed algorithm is applied to find minimax optimal designs for the logistic model, the flexible 4-parameter Hill homoscedastic model and the general nth order consecutive reaction model, and shows that it (i) produces designs that compare well with minimax D−optimal de- signs obtained from semi-infinite programming method in the literature; (ii) can be applied to semidefinite representable optimality criteria, that include the com- mon A−; E−; G−; I− and D-optimality criteria; (iii) can tackle design problems with arbitrary linear constraints on the weights; and (iv) is fast and relatively easy to use.
    [Show full text]
  • Multi-Objective Optimization of Unidirectional Non-Isolated Dc/Dcconverters
    MULTI-OBJECTIVE OPTIMIZATION OF UNIDIRECTIONAL NON-ISOLATED DC/DC CONVERTERS by Andrija Stupar A thesis submitted in conformity with the requirements for the degree of Doctor of Philosophy Graduate Department of Edward S. Rogers Department of Electrical and Computer Engineering University of Toronto © Copyright 2017 by Andrija Stupar Abstract Multi-Objective Optimization of Unidirectional Non-Isolated DC/DC Converters Andrija Stupar Doctor of Philosophy Graduate Department of Edward S. Rogers Department of Electrical and Computer Engineering University of Toronto 2017 Engineers have to fulfill multiple requirements and strive towards often competing goals while designing power electronic systems. This can be an analytically complex and computationally intensive task since the relationship between the parameters of a system’s design space is not always obvious. Furthermore, a number of possible solutions to a particular problem may exist. To find an optimal system, many different possible designs must be evaluated. Literature on power electronics optimization focuses on the modeling and design of partic- ular converters, with little thought given to the mathematical formulation of the optimization problem. Therefore, converter optimization has generally been a slow process, with exhaus- tive search, the execution time of which is exponential in the number of design variables, the prevalent approach. In this thesis, geometric programming (GP), a type of convex optimization, the execution time of which is polynomial in the number of design variables, is proposed and demonstrated as an efficient and comprehensive framework for the multi-objective optimization of non-isolated unidirectional DC/DC converters. A GP model of multilevel flying capacitor step-down convert- ers is developed and experimentally verified on a 15-to-3.3 V, 9.9 W discrete prototype, with sets of loss-volume Pareto optimal designs generated in under one minute.
    [Show full text]
  • Xpress Nonlinear Manual This Material Is the Confidential, Proprietary, and Unpublished Property of Fair Isaac Corporation
    R FICO Xpress Optimization Last update 01 May, 2018 version 33.01 FICO R Xpress Nonlinear Manual This material is the confidential, proprietary, and unpublished property of Fair Isaac Corporation. Receipt or possession of this material does not convey rights to divulge, reproduce, use, or allow others to use it without the specific written authorization of Fair Isaac Corporation and use must conform strictly to the license agreement. The information in this document is subject to change without notice. If you find any problems in this documentation, please report them to us in writing. Neither Fair Isaac Corporation nor its affiliates warrant that this documentation is error-free, nor are there any other warranties with respect to the documentation except as may be provided in the license agreement. ©1983–2018 Fair Isaac Corporation. All rights reserved. Permission to use this software and its documentation is governed by the software license agreement between the licensee and Fair Isaac Corporation (or its affiliate). Portions of the program may contain copyright of various authors and may be licensed under certain third-party licenses identified in the software, documentation, or both. In no event shall Fair Isaac Corporation or its affiliates be liable to any person for direct, indirect, special, incidental, or consequential damages, including lost profits, arising out of the use of this software and its documentation, even if Fair Isaac Corporation or its affiliates have been advised of the possibility of such damage. The rights and allocation of risk between the licensee and Fair Isaac Corporation (or its affiliates) are governed by the respective identified licenses in the software, documentation, or both.
    [Show full text]
  • Julia, My New Friend for Computing and Optimization? Pierre Haessig, Lilian Besson
    Julia, my new friend for computing and optimization? Pierre Haessig, Lilian Besson To cite this version: Pierre Haessig, Lilian Besson. Julia, my new friend for computing and optimization?. Master. France. 2018. cel-01830248 HAL Id: cel-01830248 https://hal.archives-ouvertes.fr/cel-01830248 Submitted on 4 Jul 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. « Julia, my new computing friend? » | 14 June 2018, IETR@Vannes | By: L. Besson & P. Haessig 1 « Julia, my New frieNd for computiNg aNd optimizatioN? » Intro to the Julia programming language, for MATLAB users Date: 14th of June 2018 Who: Lilian Besson & Pierre Haessig (SCEE & AUT team @ IETR / CentraleSupélec campus Rennes) « Julia, my new computing friend? » | 14 June 2018, IETR@Vannes | By: L. Besson & P. Haessig 2 AgeNda for today [30 miN] 1. What is Julia? [5 miN] 2. ComparisoN with MATLAB [5 miN] 3. Two examples of problems solved Julia [5 miN] 4. LoNger ex. oN optimizatioN with JuMP [13miN] 5. LiNks for more iNformatioN ? [2 miN] « Julia, my new computing friend? » | 14 June 2018, IETR@Vannes | By: L. Besson & P. Haessig 3 1. What is Julia ? Open-source and free programming language (MIT license) Developed since 2012 (creators: MIT researchers) Growing popularity worldwide, in research, data science, finance etc… Multi-platform: Windows, Mac OS X, GNU/Linux..
    [Show full text]
  • Solving Mixed Integer Linear and Nonlinear Problems Using the SCIP Optimization Suite
    Takustraße 7 Konrad-Zuse-Zentrum D-14195 Berlin-Dahlem fur¨ Informationstechnik Berlin Germany TIMO BERTHOLD GERALD GAMRATH AMBROS M. GLEIXNER STEFAN HEINZ THORSTEN KOCH YUJI SHINANO Solving mixed integer linear and nonlinear problems using the SCIP Optimization Suite Supported by the DFG Research Center MATHEON Mathematics for key technologies in Berlin. ZIB-Report 12-27 (July 2012) Herausgegeben vom Konrad-Zuse-Zentrum f¨urInformationstechnik Berlin Takustraße 7 D-14195 Berlin-Dahlem Telefon: 030-84185-0 Telefax: 030-84185-125 e-mail: [email protected] URL: http://www.zib.de ZIB-Report (Print) ISSN 1438-0064 ZIB-Report (Internet) ISSN 2192-7782 Solving mixed integer linear and nonlinear problems using the SCIP Optimization Suite∗ Timo Berthold Gerald Gamrath Ambros M. Gleixner Stefan Heinz Thorsten Koch Yuji Shinano Zuse Institute Berlin, Takustr. 7, 14195 Berlin, Germany, fberthold,gamrath,gleixner,heinz,koch,[email protected] July 31, 2012 Abstract This paper introduces the SCIP Optimization Suite and discusses the ca- pabilities of its three components: the modeling language Zimpl, the linear programming solver SoPlex, and the constraint integer programming frame- work SCIP. We explain how these can be used in concert to model and solve challenging mixed integer linear and nonlinear optimization problems. SCIP is currently one of the fastest non-commercial MIP and MINLP solvers. We demonstrate the usage of Zimpl, SCIP, and SoPlex by selected examples, give an overview of available interfaces, and outline plans for future development. ∗A Japanese translation of this paper will be published in the Proceedings of the 24th RAMP Symposium held at Tohoku University, Miyagi, Japan, 27{28 September 2012, see http://orsj.or.
    [Show full text]
  • [20Pt]Algorithms for Constrained Optimization: [ 5Pt]
    SOL Optimization 1970s 1980s 1990s 2000s 2010s Summary 2020s Algorithms for Constrained Optimization: The Benefits of General-purpose Software Michael Saunders MS&E and ICME, Stanford University California, USA 3rd AI+IoT Business Conference Shenzhen, China, April 25, 2019 Optimization Software 3rd AI+IoT Business Conference, Shenzhen, April 25, 2019 1/39 SOL Optimization 1970s 1980s 1990s 2000s 2010s Summary 2020s SOL Systems Optimization Laboratory George Dantzig, Stanford University, 1974 Inventor of the Simplex Method Father of linear programming Large-scale optimization: Algorithms, software, applications Optimization Software 3rd AI+IoT Business Conference, Shenzhen, April 25, 2019 2/39 SOL Optimization 1970s 1980s 1990s 2000s 2010s Summary 2020s SOL history 1974 Dantzig and Cottle start SOL 1974{78 John Tomlin, LP/MIP expert 1974{2005 Alan Manne, nonlinear economic models 1975{76 MS, MINOS first version 1979{87 Philip Gill, Walter Murray, MS, Margaret Wright (Gang of 4!) 1989{ Gerd Infanger, stochastic optimization 1979{ Walter Murray, MS, many students 2002{ Yinyu Ye, optimization algorithms, especially interior methods This week! UC Berkeley opened George B. Dantzig Auditorium Optimization Software 3rd AI+IoT Business Conference, Shenzhen, April 25, 2019 3/39 SOL Optimization 1970s 1980s 1990s 2000s 2010s Summary 2020s Optimization problems Minimize an objective function subject to constraints: 0 x 1 min '(x) st ` ≤ @ Ax A ≤ u c(x) x variables 0 1 A matrix c1(x) B . C c(x) nonlinear functions @ . A c (x) `; u bounds m Optimization
    [Show full text]
  • Pysp: Modeling and Solving Stochastic Programs in Python
    Noname manuscript No. (will be inserted by the editor) PySP: Modeling and Solving Stochastic Programs in Python Jean-Paul Watson · David L. Woodruff · William E. Hart Received: September 6, 2010. Abstract Although stochastic programming is a powerful tool for modeling decision- making under uncertainty, various impediments have historically prevented its wide- spread use. One key factor involves the ability of non-specialists to easily express stochastic programming problems as extensions of deterministic models, which are often formulated first. A second key factor relates to the difficulty of solving stochastic programming models, particularly the general mixed-integer, multi-stage case. Intri- cate, configurable, and parallel decomposition strategies are frequently required to achieve tractable run-times. We simultaneously address both of these factors in our PySP software package, which is part of the COIN-OR Coopr open-source Python project for optimization. To formulate a stochastic program in PySP, the user speci- fies both the deterministic base model and the scenario tree with associated uncertain parameters in the Pyomo open-source algebraic modeling language. Given these two models, PySP provides two paths for solution of the corresponding stochastic program. The first alternative involves writing the extensive form and invoking a standard deter- ministic (mixed-integer) solver. For more complex stochastic programs, we provide an implementation of Rockafellar and Wets’ Progressive Hedging algorithm. Our particu- lar focus is on the use of Progressive Hedging as an effective heuristic for approximating Jean-Paul Watson Sandia National Laboratories Discrete Math and Complex Systems Department PO Box 5800, MS 1318 Albuquerque, NM 87185-1318 E-mail: [email protected] David L.
    [Show full text]
  • TOMLAB –Unique Features for Optimization in MATLAB
    TOMLAB –Unique Features for Optimization in MATLAB Bad Honnef, Germany October 15, 2004 Kenneth Holmström Tomlab Optimization AB Västerås, Sweden [email protected] ´ Professor in Optimization Department of Mathematics and Physics Mälardalen University, Sweden http://tomlab.biz Outline of the talk • The TOMLAB Optimization Environment – Background and history – Technology available • Optimization in TOMLAB • Tests on customer supplied large-scale optimization examples • Customer cases, embedded solutions, consultant work • Business perspective • Box-bounded global non-convex optimization • Summary http://tomlab.biz Background • MATLAB – a high-level language for mathematical calculations, distributed by MathWorks Inc. • Matlab can be extended by Toolboxes that adds to the features in the software, e.g.: finance, statistics, control, and optimization • Why develop the TOMLAB Optimization Environment? – A uniform approach to optimization didn’t exist in MATLAB – Good optimization solvers were missing – Large-scale optimization was non-existent in MATLAB – Other toolboxes needed robust and fast optimization – Technical advantages from the MATLAB languages • Fast algorithm development and modeling of applied optimization problems • Many built in functions (ODE, linear algebra, …) • GUIdevelopmentfast • Interfaceable with C, Fortran, Java code http://tomlab.biz History of Tomlab • President and founder: Professor Kenneth Holmström • The company founded 1986, Development started 1989 • Two toolboxes NLPLIB och OPERA by 1995 • Integrated format for optimization 1996 • TOMLAB introduced, ISMP97 in Lausanne 1997 • TOMLAB v1.0 distributed for free until summer 1999 • TOMLAB v2.0 first commercial version fall 1999 • TOMLAB starting sales from web site March 2000 • TOMLAB v3.0 expanded with external solvers /SOL spring 2001 • Dash Optimization Ltd’s XpressMP added in TOMLAB fall 2001 • Tomlab Optimization Inc.
    [Show full text]
  • Specifying “Logical” Conditions in AMPL Optimization Models
    Specifying “Logical” Conditions in AMPL Optimization Models Robert Fourer AMPL Optimization www.ampl.com — 773-336-AMPL INFORMS Annual Meeting Phoenix, Arizona — 14-17 October 2012 Session SA15, Software Demonstrations Robert Fourer, Logical Conditions in AMPL INFORMS Annual Meeting — 14-17 Oct 2012 — Session SA15, Software Demonstrations 1 New and Forthcoming Developments in the AMPL Modeling Language and System Optimization modelers are often stymied by the complications of converting problem logic into algebraic constraints suitable for solvers. The AMPL modeling language thus allows various logical conditions to be described directly. Additionally a new interface to the ILOG CP solver handles logic in a natural way not requiring conventional transformations. Robert Fourer, Logical Conditions in AMPL INFORMS Annual Meeting — 14-17 Oct 2012 — Session SA15, Software Demonstrations 2 AMPL News Free AMPL book chapters AMPL for Courses Extended function library Extended support for “logical” conditions AMPL driver for CPLEX Opt Studio “Concert” C++ interface Support for ILOG CP constraint programming solver Support for “logical” constraints in CPLEX INFORMS Impact Prize to . Originators of AIMMS, AMPL, GAMS, LINDO, MPL Awards presented Sunday 8:30-9:45, Conv Ctr West 101 Doors close 8:45! Robert Fourer, Logical Conditions in AMPL INFORMS Annual Meeting — 14-17 Oct 2012 — Session SA15, Software Demonstrations 3 AMPL Book Chapters now free for download www.ampl.com/BOOK/download.html Bound copies remain available purchase from usual
    [Show full text]
  • Using COIN-OR to Solve the Uncapacitated Facility Location Problem
    The Uncapacitated Facility Location Problem Developing a Solver Additional Resources Using COIN-OR to Solve the Uncapacitated Facility Location Problem Ted Ralphs1 Matthew Saltzman 2 Matthew Galati 3 1COR@L Lab Department of Industrial and Systems Engineering Lehigh University 2Department of Mathematical Sciences Clemson University 3Analytical Solutions SAS Institute EURO XXI, July 4, 2006 Ted Ralphs, Matthew Saltzman , Matthew Galati Using COIN-OR to Solve the Uncapacitated Facility Location Problem The Uncapacitated Facility Location Problem Developing a Solver Additional Resources Outline 1 The Uncapacitated Facility Location Problem UFL Formulation Cutting Planes 2 Developing a Solver The ufl Class COIN Tools Putting It All Together 3 Additional Resources Ted Ralphs, Matthew Saltzman , Matthew Galati Using COIN-OR to Solve the Uncapacitated Facility Location Problem The Uncapacitated Facility Location Problem UFL Formulation Developing a Solver Cutting Planes Additional Resources Input Data The following are the input data needed to describe an instance of the uncapacitated facility location problem (UFL): Data a set of depots N = {1, ..., n}, a set of clients M = {1, ..., m}, the unit transportation cost cij to service client i from depot j, the fixed cost fj for using depot j Variables xij is the amount of the demand for client i satisfied from depot j yj is 1 if the depot is used, 0 otherwise Ted Ralphs, Matthew Saltzman , Matthew Galati Using COIN-OR to Solve the Uncapacitated Facility Location Problem The Uncapacitated Facility
    [Show full text]
  • Pyomo Documentation Release 5.6.2.Dev0
    Pyomo Documentation Release 5.6.2.dev0 Pyomo Feb 06, 2019 Contents 1 Installation 3 1.1 Using CONDA..............................................3 1.2 Using PIP.................................................3 2 Citing Pyomo 5 2.1 Pyomo..................................................5 2.2 PySP...................................................5 3 Pyomo Overview 7 3.1 Mathematical Modeling.........................................7 3.2 Overview of Modeling Components and Processes...........................9 3.3 Abstract Versus Concrete Models....................................9 3.4 Simple Models.............................................. 10 4 Pyomo Modeling Components 17 4.1 Sets.................................................... 17 4.2 Parameters................................................ 23 4.3 Variables................................................. 24 4.4 Objectives................................................ 25 4.5 Constraints................................................ 26 4.6 Expressions................................................ 26 4.7 Suffixes.................................................. 31 5 Solving Pyomo Models 41 5.1 Solving ConcreteModels......................................... 41 5.2 Solving AbstractModels......................................... 41 5.3 pyomo solve Command....................................... 41 5.4 Supported Solvers............................................ 42 6 Working with Pyomo Models 43 6.1 Repeated Solves............................................. 43 6.2 Changing
    [Show full text]