Process Optimization
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
How to Use Your Favorite MIP Solver: Modeling, Solving, Cannibalizing
How to use your favorite MIP Solver: modeling, solving, cannibalizing Andrea Lodi University of Bologna, Italy [email protected] January-February, 2012 @ Universit¨atWien A. Lodi, How to use your favorite MIP Solver Setting • We consider a general Mixed Integer Program in the form: T maxfc x : Ax ≤ b; x ≥ 0; xj 2 Z; 8j 2 Ig (1) where matrix A does not have a special structure. A. Lodi, How to use your favorite MIP Solver 1 Setting • We consider a general Mixed Integer Program in the form: T maxfc x : Ax ≤ b; x ≥ 0; xj 2 Z; 8j 2 Ig (1) where matrix A does not have a special structure. • Thus, the problem is solved through branch-and-bound and the bounds are computed by iteratively solving the LP relaxations through a general-purpose LP solver. A. Lodi, How to use your favorite MIP Solver 1 Setting • We consider a general Mixed Integer Program in the form: T maxfc x : Ax ≤ b; x ≥ 0; xj 2 Z; 8j 2 Ig (1) where matrix A does not have a special structure. • Thus, the problem is solved through branch-and-bound and the bounds are computed by iteratively solving the LP relaxations through a general-purpose LP solver. • The course basically covers the MIP but we will try to discuss when possible how crucial is the LP component (the engine), and how much the whole framework is built on top the capability of effectively solving LPs. • Roughly speaking, using the LP computation as a tool, MIP solvers integrate the branch-and-bound and the cutting plane algorithms through variations of the general branch-and-cut scheme [Padberg & Rinaldi 1987] developed in the context of the Traveling Salesman Problem (TSP). -
12. Coordinate Descent Methods
EE 546, Univ of Washington, Spring 2014 12. Coordinate descent methods theoretical justifications • randomized coordinate descent method • minimizing composite objectives • accelerated coordinate descent method • Coordinate descent methods 12–1 Notations consider smooth unconstrained minimization problem: minimize f(x) x RN ∈ n coordinate blocks: x =(x ,...,x ) with x RNi and N = N • 1 n i ∈ i=1 i more generally, partition with a permutation matrix: U =[PU1 Un] • ··· n T xi = Ui x, x = Uixi i=1 X blocks of gradient: • f(x) = U T f(x) ∇i i ∇ coordinate update: • x+ = x tU f(x) − i∇i Coordinate descent methods 12–2 (Block) coordinate descent choose x(0) Rn, and iterate for k =0, 1, 2,... ∈ 1. choose coordinate i(k) 2. update x(k+1) = x(k) t U f(x(k)) − k ik∇ik among the first schemes for solving smooth unconstrained problems • cyclic or round-Robin: difficult to analyze convergence • mostly local convergence results for particular classes of problems • does it really work (better than full gradient method)? • Coordinate descent methods 12–3 Steepest coordinate descent choose x(0) Rn, and iterate for k =0, 1, 2,... ∈ (k) 1. choose i(k) = argmax if(x ) 2 i 1,...,n k∇ k ∈{ } 2. update x(k+1) = x(k) t U f(x(k)) − k i(k)∇i(k) assumptions f(x) is block-wise Lipschitz continuous • ∇ f(x + U v) f(x) L v , i =1,...,n k∇i i −∇i k2 ≤ ik k2 f has bounded sub-level set, in particular, define • ⋆ R(x) = max max y x 2 : f(y) f(x) y x⋆ X⋆ k − k ≤ ∈ Coordinate descent methods 12–4 Analysis for constant step size quadratic upper bound due to block coordinate-wise -
The Gradient Sampling Methodology
The Gradient Sampling Methodology James V. Burke∗ Frank E. Curtisy Adrian S. Lewisz Michael L. Overtonx January 14, 2019 1 Introduction in particular, this leads to calling the negative gra- dient, namely, −∇f(x), the direction of steepest de- The principal methodology for minimizing a smooth scent for f at x. However, when f is not differentiable function is the steepest descent (gradient) method. near x, one finds that following the negative gradient One way to extend this methodology to the minimiza- direction might offer only a small amount of decrease tion of a nonsmooth function involves approximating in f; indeed, obtaining decrease from x along −∇f(x) subdifferentials through the random sampling of gra- may be possible only with a very small stepsize. The dients. This approach, known as gradient sampling GS methodology is based on the idea of stabilizing (GS), gained a solid theoretical foundation about a this definition of steepest descent by instead finding decade ago [BLO05, Kiw07], and has developed into a a direction to approximately solve comprehensive methodology for handling nonsmooth, potentially nonconvex functions in the context of op- min max gT d; (2) kdk2≤1 g2@¯ f(x) timization algorithms. In this article, we summarize the foundations of the GS methodology, provide an ¯ where @f(x) is the -subdifferential of f at x [Gol77]. overview of the enhancements and extensions to it To understand the context of this idea, recall that that have developed over the past decade, and high- the (Clarke) subdifferential of a locally Lipschitz f light some interesting open questions related to GS. -
Notes 1: Introduction to Optimization Models
Notes 1: Introduction to Optimization Models IND E 599 September 29, 2010 IND E 599 Notes 1 Slide 1 Course Objectives I Survey of optimization models and formulations, with focus on modeling, not on algorithms I Include a variety of applications, such as, industrial, mechanical, civil and electrical engineering, financial optimization models, health care systems, environmental ecology, and forestry I Include many types of optimization models, such as, linear programming, integer programming, quadratic assignment problem, nonlinear convex problems and black-box models I Include many common formulations, such as, facility location, vehicle routing, job shop scheduling, flow shop scheduling, production scheduling (min make span, min max lateness), knapsack/multi-knapsack, traveling salesman, capacitated assignment problem, set covering/packing, network flow, shortest path, and max flow. IND E 599 Notes 1 Slide 2 Tentative Topics Each topic is an introduction to what could be a complete course: 1. basic linear models (LP) with sensitivity analysis 2. integer models (IP), such as the assignment problem, knapsack problem and the traveling salesman problem 3. mixed integer formulations 4. quadratic assignment problems 5. include uncertainty with chance-constraints, stochastic programming scenario-based formulations, and robust optimization 6. multi-objective formulations 7. nonlinear formulations, as often found in engineering design 8. brief introduction to constraint logic programming 9. brief introduction to dynamic programming IND E 599 Notes 1 Slide 3 Computer Software I Catalyst Tools (https://catalyst.uw.edu/) I AIMMS - optimization software (http://www.aimms.com/) Ming Fang - AIMMS software consultant IND E 599 Notes 1 Slide 4 What is Mathematical Programming? Mathematical programming refers to \programming" as a \planning" activity: as in I linear programming (LP) I integer programming (IP) I mixed integer linear programming (MILP) I non-linear programming (NLP) \Optimization" is becoming more common, e.g. -
A New Algorithm of Nonlinear Conjugate Gradient Method with Strong Convergence*
Volume 27, N. 1, pp. 93–106, 2008 Copyright © 2008 SBMAC ISSN 0101-8205 www.scielo.br/cam A new algorithm of nonlinear conjugate gradient method with strong convergence* ZHEN-JUN SHI1,2 and JINHUA GUO2 1College of Operations Research and Management, Qufu Normal University Rizhao, Sahndong 276826, P.R. China 2Department of Computer and Information Science, University of Michigan Dearborn, Michigan 48128-1491, USA E-mails: [email protected]; [email protected] / [email protected] Abstract. The nonlinear conjugate gradient method is a very useful technique for solving large scale minimization problems and has wide applications in many fields. In this paper, we present a new algorithm of nonlinear conjugate gradient method with strong convergence for unconstrained minimization problems. The new algorithm can generate an adequate trust region radius automatically at each iteration and has global convergence and linear convergence rate under some mild conditions. Numerical results show that the new algorithm is efficient in practical computation and superior to other similar methods in many situations. Mathematical subject classification: 90C30, 65K05, 49M37. Key words: unconstrained optimization, nonlinear conjugate gradient method, global conver- gence, linear convergence rate. 1 Introduction Consider an unconstrained minimization problem min f (x), x ∈ Rn, (1) where Rn is an n-dimensional Euclidean space and f : Rn −→ R is a continu- ously differentiable function. #724/07. Received: 10/V/07. Accepted: 24/IX/07. *The work was supported in part by NSF CNS-0521142, USA. 94 A NEW ALGORITHM OF NONLINEAR CONJUGATE GRADIENT METHOD When n is very large (for example, n > 106) the related problem is called large scale minimization problem. -
Co-Optimization of Transmission and Other Supply Resources
Co-optimization of Transmission and Other Supply Resources September, 2013 Illinois Institute of Technology Iowa State University Johns Hopkins University Purdue University West Virginia University For EISPC and NARUC Funded by the U.S Department of Energy Co-optimization of Transmission and Other Supply Resources prepared for Eastern Interconnection States’ Planning Council and National Association of Regulatory Utility Commissioners prepared by Dr. Andrew (Lu) Liu, Purdue University Dr. Benjamin H. Hobbs and Jonathan Ho, Johns Hopkins University Dr. James D. McCalley and Venkat Krishnan, Iowa State University Dr. Mohammad Shahidehpour, Illinois Institute of Technology Dr. Qipeng P. Zheng, University of Central Florida Acknowledgement: This material is based upon work supported by the Department of Energy, National Energy Technology Laboratory, under Award Number DE-OE0000316. The project team would like to thank Bob Pauley, Doug Gotham and Stan Hadley as well as the EISPC and NARUC organizations, for their support throughout this project. Patrick Sullivan of NREL provided crucial perspectives, advice, and resources during the project. We would also like to acknowledge inputs on earlier versions of our work by members of NARUC’s Studies and Whitepaper Workgroup. The earlier versions as well benefited from the editorial assistance provided by Jingjie Xiao, a former PhD student with the School of Industrial Engineering at Purdue University, and her generous efforts are greatly appreciated. In addition, the team gratefully acknowledge the generous and timely assistance of several organizations that have provided important background information on transmission planning, including but not limited to: Parveen Baig, Iowa Utilities Board Jaeseok Choi, Gyeongsang National University, South Korea Bruce Fardanesh, New York Power Authority Flora Flydt, ATC Ian Grant, TVA Bruce Hamilton, Smart Grid Network Ben Harrison, Duke Energy Michael Henderson, ISO New England Len Januzik, Quanta Technology David Kelley, SPP Raymond Kershaw, ITC Holdings Corp. -
MODELING LANGUAGES in MATHEMATICAL OPTIMIZATION Applied Optimization Volume 88
MODELING LANGUAGES IN MATHEMATICAL OPTIMIZATION Applied Optimization Volume 88 Series Editors: Panos M. Pardalos University o/Florida, U.S.A. Donald W. Hearn University o/Florida, U.S.A. MODELING LANGUAGES IN MATHEMATICAL OPTIMIZATION Edited by JOSEF KALLRATH BASF AG, GVC/S (Scientific Computing), 0-67056 Ludwigshafen, Germany Dept. of Astronomy, Univ. of Florida, Gainesville, FL 32611 Kluwer Academic Publishers Boston/DordrechtiLondon Distributors for North, Central and South America: Kluwer Academic Publishers 101 Philip Drive Assinippi Park Norwell, Massachusetts 02061 USA Telephone (781) 871-6600 Fax (781) 871-6528 E-Mail <[email protected]> Distributors for all other countries: Kluwer Academic Publishers Group Post Office Box 322 3300 AlI Dordrecht, THE NETHERLANDS Telephone 31 78 6576 000 Fax 31 786576474 E-Mail <[email protected]> .t Electronic Services <http://www.wkap.nl> Library of Congress Cataloging-in-Publication Kallrath, Josef Modeling Languages in Mathematical Optimization ISBN-13: 978-1-4613-7945-4 e-ISBN-13:978-1-4613 -0215 - 5 DOl: 10.1007/978-1-4613-0215-5 Copyright © 2004 by Kluwer Academic Publishers Softcover reprint of the hardcover 1st edition 2004 All rights reserved. No part ofthis pUblication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photo-copying, microfilming, recording, or otherwise, without the prior written permission ofthe publisher, with the exception of any material supplied specifically for the purpose of being entered and executed on a computer system, for exclusive use by the purchaser of the work. Permissions for books published in the USA: P"§.;r.m.i.§J?i..QD.§.@:w.k~p" ..,.g.Qm Permissions for books published in Europe: [email protected] Printed on acid-free paper. -
Download Classic LINDO User's Manual
LINDO User’s Manual LINDO Systems, Inc. 1415 North Dayton Street, Chicago, Illinois 60622 Phone: (312)988-7422 Fax: (312)988-9065 E-mail: [email protected] WWW: http://www.lindo.com COPYRIGHT LINDO software and its related documentation are copyrighted. You may not copy the LINDO software or related documentation except in the manner authorized in the related documentation or with the written permission of LINDO systems, Inc. TRADEMARKS LINGO is a trademark, and LINDO and What’sBest! are registered trademarks, of LINDO Systems, Inc. Other product and company names mentioned herein are the property of their respective owners. DISCLAIMER LINDO Systems, Inc. warrants that on the date of receipt of your payment, the disk enclosed in the disk envelope contains an accurate reproduction of the LINDO software and that the copy of the related documentation is accurately reproduced. Due to the inherent complexity of computer programs and computer models, the LINDO software may not be completely free of errors. You are advised to verify your answers before basing decisions on them. NEITHER LINDO SYSTEMS, INC. NOR ANYONE ELSE ASSOCIATED IN THE CREATION, PRODUCTION, OR DISTRIBUTION OF THE LINDO SOFTWARE MAKES ANY OTHER EXPRESSED WARRANTIES REGARDING THE DISKS OR DOCUMENTATION AND MAKES NO WARRANTIES AT ALL, EITHER EXPRESSED OR IMPLIED, REGARDING THE LINDO SOFTWARE, INCLUDING THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE OR OTHERWISE. Further, LINDO Systems, Inc. reserves the right to revise this software and related documentation and make changes to the content hereof without obligation to notify any person of such revisions or changes. -
Using a New Nonlinear Gradient Method for Solving Large Scale Convex Optimization Problems with an Application on Arabic Medical Text
Using a New Nonlinear Gradient Method for Solving Large Scale Convex Optimization Problems with an Application on Arabic Medical Text Jaafar Hammouda*, Ali Eisab, Natalia Dobrenkoa, Natalia Gusarovaa aITMO University, St. Petersburg, Russia *[email protected] bAleppo University, Aleppo, Syria Abstract Gradient methods have applications in multiple fields, including signal processing, image processing, and dynamic systems. In this paper, we present a nonlinear gradient method for solving convex supra-quadratic functions by developing the search direction, that done by hybridizing between the two conjugate coefficients HRM [2] and NHS [1]. The numerical results proved the effectiveness of the presented method by applying it to solve standard problems and reaching the exact solution if the objective function is quadratic convex. Also presented in this article, an application to the problem of named entities in the Arabic medical language, as it proved the stability of the proposed method and its efficiency in terms of execution time. Keywords: Convex optimization; Gradient methods; Named entity recognition; Arabic; e-Health. 1. Introduction Several nonlinear conjugate gradient methods have been presented for solving high-dimensional unconstrained optimization problems which are given as [3]: min f(x) ; x∈Rn (1) To solve problem (1), we start with the following iterative relationship: 푥(푘+1) = 푥푘 + 훼푘 푑푘, 푘 = 0,1,2, (2) where α푘 > 0 is a step size, that is calculated by strong Wolfe-Powell’s conditions [4] 푇 푓(푥푘 + 훼푘푑푘) ≤ 푓(푥푘) + 훿훼푘푔푘 푑푘, (3) 푇 푇 |푔(푥푘 + 훼푘푑푘) 푑푘| ≤ 휎|푔푘 푑푘| where 0 < 훿 < 휎 < 1 푑푘 is the search direction that is computed as follow [3] −푔푘 , if 푘 = 0 (4) 푑푘+1 = { −푔푘+1 + 훽푘푑푘 , if 푘 ≥ 1 Where 푔푘 = 푔(푥푘) = ∇푓(푥푘) represent the gradient vector for 푓(푥) at the point 푥푘 , 훽푘 ∈ ℝ is known as CG coefficient that characterizes different CG methods. -
Open Source and Free Software
WP. 24 ENGLISH ONLY UNITED NATIONS STATISTICAL COMMISSION and EUROPEAN COMMISSION ECONOMIC COMMISSION FOR EUROPE STATISTICAL OFFICE OF THE CONFERENCE OF EUROPEAN STATISTICIANS EUROPEAN COMMUNITIES (EUROSTAT) Joint UNECE/Eurostat work session on statistical data confidentiality (Bilbao, Spain, 2-4 December 2009) Topic (iv): Tools and software improvements ON OPEN SOURCE SOFTWARE FOR STATISTICAL DISCLOSURE LIMITATION Invited Paper Prepared by Juan José Salazar González, University of La Laguna, Spain On open source software for Statistical Disclosure Limitation Juan José Salazar González * * Department of Statistics, Operations Research and Computer Science, University of La Laguna, 38271 Tenerife, Spain, e-mail: [email protected] Abstract: Much effort has been done in the last years to design, analyze, implement and compare different methodologies to ensure confidentiality during data publication. The knowledge of these methods is public, but the practical implementations are subject to different license constraints. This paper summarizes some concepts on free and open source software, and gives some light on the convenience of using these schemes when building automatic tools in Statistical Disclosure Limitation. Our results are based on some computational experiments comparing different mathematical programming solvers when applying controlled rounding to tabular data. 1 Introduction Statistical agencies must guarantee confidentiality of data respondent by using the best of modern technology. Today data snoopers may have powerful computers, thus statistical agencies need sophisticated computer programs to ensure that released information is protected against attackers. Recent research has designed different optimization techniques that ensure protection. Unfortunately the number of users of these techniques is quite small (mainly national and regional statistical agencies), thus there is not enough market to stimulate the development of several implementations, all competing for being the best. -
Chipps Framework Applications Results and Conclusions
Introduction The CHiPPS Framework Applications Results and Conclusions The COIN-OR High-Performance Parallel Search Framework (CHiPPS) TED RALPHS LEHIGH UNIVERSITY YAN XU SAS INSTITUTE CPAIOR, June 15, 2010 Thanks: Work supported in part by the National Science Foundation and IBM Ralphs& Xu CHiPPS Introduction The CHiPPS Framework Applications Results and Conclusions Outline 1 Introduction Tree Search Algorithms Parallel Computing Previous Work 2 The CHiPPS Framework Introduction ALPS: Abstract Library For Parallel Search BiCePS: Branch, Constrain, and Price Software BLIS: BiCePS Linear Integer Solver 3 Applications Knapsack Problem Vehicle Routing 4 Results and Conclusions Ralphs& Xu CHiPPS Introduction Tree Search Algorithms The CHiPPS Framework Parallel Computing Applications Previous Work Results and Conclusions Tree Search Algorithms Tree search algorithms systematically search the nodes of an acyclic graph for certain goal nodes. Root Initial State Goal State Tree search algorithms have been applied in many areas such as Constraint satisfaction, Game search, Constraint Programming, and Mathematical programming. Ralphs& Xu CHiPPS Introduction Tree Search Algorithms The CHiPPS Framework Parallel Computing Applications Previous Work Results and Conclusions Elements of Tree Search Algorithms A generic tree search algorithm consists of the following elements: Generic Tree Search Algorithm Processing method: Is this a goal node? Fathoming rule: Can node can be fathomed? Branching method: What are the successors of this node? Search strategy: What should we work on next? The algorithm consists of choosing a candidate node, processing it, and either fathoming or branching. During the course of the search, various information (knowledge) is generated and can be used to guide the search. Ralphs& Xu CHiPPS Introduction Tree Search Algorithms The CHiPPS Framework Parallel Computing Applications Previous Work Results and Conclusions Parallelizing Tree Search Algorithms In general, the search tree can be very large. -
Conditional Gradient Method for Multiobjective Optimization
Conditional gradient method for multiobjective optimization P. B. Assun¸c~ao ∗ O. P. Ferreira ∗ L. F. Prudente ∗ April 7, 2020 Abstract: We analyze the conditional gradient method, also known as Frank-Wolfe method, for constrained multiobjective optimization. The constraint set is assumed to be convex and com- pact, and the objectives functions are assumed to be continuously differentiable. The method is considered with different strategies for obtaining the step sizes. Asymptotic convergence proper- ties and iteration-complexity bounds with and without convexity assumptions on the objective functions are stablished. Numerical experiments are provided to illustrate the effectiveness of the method and certify the obtained theoretical results. Keywords: Conditional gradient method; multiobjective optimization; Pareto optimality; con- strained optimization problem 1 Introduction In the constrained multiobjective optimization, we seek to simultaneously minimize several objec- tive functions on a set C. One strategy for solving multiobjective problems that has become very popular consists in the extension of methods for scalar-valued to multiobjective-valued optimiza- tion, instead of using scalarization approaches [23]. As far as we know, this strategy originated from the work [14] in which the authors proposed the steepest descent methods for multiobjec- tive optimization. Since of then, new properties related to this method have been discovered and several variants of it have been considered, see for example [4,5,16,19,20,26,27]. In recent years, there has been a significant increase in the number of papers addressing concepts, techniques, and methods for multiobjective optimization, see for example [6, 9, 13, 15, 25, 41, 42, 45, 46, 48, 53, 54].