Dynamic Matrix Algorithms and Applications in Convex and Combinatorial Optimization

Total Page:16

File Type:pdf, Size:1020Kb

Dynamic Matrix Algorithms and Applications in Convex and Combinatorial Optimization kth royal institute of technology Doctoral Thesis in Computer Science Dynamic Matrix Algorithms and Applications in Convex and Combinatorial Optimization JAN VAN DEN BRAND Stockholm, Sweden 2021 Dynamic Matrix Algorithms and Applications in Convex and Combinatorial Optimization JAN VAN DEN BRAND Academic Dissertation which, with due permission of the KTH Royal Institute of Technology, is submitted for public defence for the Degree of Doctor of Philosophy on Wednesday the 9th of June 2021, at 3:00 p.m. in F3, Lindstedsvägen 26, Stockholm. Doctoral Thesis in Computer Science KTH Royal Institute of Technology Stockholm, Sweden 2021 © Jan van den Brand © Co-Authors: Yin Tat Lee, Yang P. Liu, Danupon Nanongkai, Richard Peng, Thatchaphol Saranurak, Aaron Sidford, Di Wang, Zhao Song ISBN 978-91-7873-867-0 TRITA-EECS-AVL-2021:31 Printed by: Universitetsservice US-AB, Sweden 2021 i Abstract Dynamic algorithms are used to efficiently maintain solutions to problems where the input undergoes some changes. This thesis studies dynamic algo- rithms that maintain solutions to linear algebra problems and we explore their applications and implications for dynamic graphs and optimization problems. Dynamic graph algorithms maintain properties of changing graphs, such as the distances in a graph that undergoes edge deletions and insertions. The main question is how to maintain the information without recomputing the solution from scratch whenever the graph changes. If maintaining the infor- mation without trivial recomputation is possible, the next natural question is how quickly the information can be maintained. This thesis makes progress on both questions: (i) We construct the first non-trivial fully dynamic graph algorithms for single-source shortest paths, diameter and other problems. This answers open questions stated in, e.g., [Demetrescu-Italiano’04]. (ii) We obtain matching upper and conditional lower bounds for the com- plexity of maintaining reachability, maximum matching, directed cycle detec- tion and many other graph properties. This settles the complexity for these problems and answers an open problem stated in [Abboud-V.Williams’14]. We get these results by reducing the dynamic graph problems to dynamic linear algebra problems for which we develop new algorithms. At the same time, conditional lower bounds for the dynamic graph problems thus imply lower bounds for dynamic linear algebra problems as well. We apply the developed techniques for dynamic linear algebra to algo- rithms for linear programs and obtain optimal (i.e. nearly-linear time) al- gorithms for dense instances of linear programs, Markov decision processes, linear L1 regression, and graph specific special cases thereof such as bipartite matching, minimum-cost flow, and (negative weight) shortest paths. For bi- partite matching on dense graphs, this is the first improvement since the clas- sic algorithms by [Dinic’70;Hopcroft-Karp’71;Karzanov’73;Ibarra-Moran’81]. The results are obtained by using that algorithms (i.e. interior point meth- ods) for these problems are iterative and must repeatedly solve linear systems and other linear algebra problems. By using techniques from dynamic lin- ear algebra (i.e. dynamic matrix algorithms), we are able to maintain the solution to these subproblems, reducing the time required per iteration. The construction of our algorithms relies on a joint analysis of the iterative algo- rithm and the dynamic matrix algorithms. On one hand, we develop robust interior point methods which are able to handle relaxations and approxi- mations to the linear algebra subroutines. On other hand, we develop fast dynamic matrix algorithms that are able to maintain the solution to these relaxed subproblems efficiently. Keywords: Dynamic Algorithm, Data Structure, Optimization, Linear Program, Bipartite Matching, Shortest Path, Maximum Flow, Minimum Cost Flow, Diameter ii Sammanfattning Dynamiska grafalgoritmer uppr¨atth˚alleregenskaper f¨or att ¨andra diagram, t.ex. avst˚anden i en graf som genomg˚arkantraderingar och ins¨attningar. Hu- vudfr˚agan ¨ar hur man uppr¨atth˚allerinformationen utan att ber¨akna l¨osningen fr˚angrunden n¨ar grafen ¨andras. Om det ¨ar m¨ojligt att bibeh˚allainformationen utan trivial omber¨akning ¨ar n¨asta naturliga fr˚agahur snabbt informationen kan uppr¨atth˚allas.Denna avhandling g¨or framsteg i b˚adafr˚agorna: (i) Vi ger de f¨orsta icke-triviala dynamiska grafalgoritmerna f¨or kortaste v¨agar, diameter och andra problem med en k¨alla. Detta besvarar ¨oppna fr˚agor fr˚ant.ex. [Demetrescu-Italiano’04]. (ii) Vi ger matchande ¨ovre och villkorade undre gr¨anser f¨or komplexiteten i att uppr¨atth˚allan˚abarhet, maximal matchning, riktad cykeldetektering och m˚angaandra grafegenskaper. Detta l¨oser komplexiteten f¨or dessa problem och svarar p˚a ¨oppna problem som anges i [Abboud-V.Williams’14]. Vi uppn˚ardessa resultat genom att reducera de dynamiska grafproble- men till dynamiska linj¨ar algebra-problem som vi utvecklar nya algoritmer f¨or. Samtidigt inneb¨ar villkorade undre gr¨anser f¨or dynamiska grafproblem s˚alundaundre gr¨anser f¨or dynamiska linj¨ara algebraproblem ocks˚a. Vi till¨ampar de utvecklade teknikerna f¨or dynamisk linj¨ar algebra p˚aal- goritmer f¨or linj¨ara program och uppn˚aroptimala (dvs. n¨astan linj¨ar tid) al- goritmer f¨or t¨ata instanser av linj¨ara program, MDP, linj¨ar L1-regression, och graf-specifika specialfall som bipartit matchning, fl¨ode av minimal kostnad, kortaste v¨agar (med negativa vikter). F¨or bipartit matchning i t¨ata grafer ¨ar detta den f¨orsta f¨orb¨attringen sedan de klassiska algoritmerna av [Dinic’70; Hopcroft-Karp’71; Karzanov’73; Ibarra-Moran’81]. Resultaten erh˚allsmed hj¨alp av att algoritmer (dvs. inrepunkts-metoder) f¨or dessa problem ¨ar iterativa och m˚aste l¨osa linj¨ara system och andra linj¨ara algebra-problem upprepade g˚anger.Genom att anv¨anda tekniker fr˚andyna- misk linj¨ar algebra kan vi bibeh˚allal¨osningen p˚adessa delproblem, vilket minskar den tid som kr¨avs per iteration. Konstruktionen av v˚araalgoritmer bygger p˚aen f¨orenad analys av den iterativa algoritmen och de dynamiska linj¨ara algebraalgoritmerna. A˚ ena sidan utvecklar vi robusta inre punktme- toder som kan hantera approximativa l¨osningar till de linj¨ara algebrasubruti- nerna. A˚ andra sidan utvecklar vi snabba dynamiska linj¨ara algebraalgoritmer som kan bibeh˚allal¨osningen p˚adessa approximativa delproblem effektivt. Acknowledgement My time at KTH was one of the most enjoyable stages of life so far. Clearly, the most influence during that time had my advisor Danupon Nanongkai and I am deeply grateful for his support. He provided me with a lot of freedom, e.g. I was able to select my own research questions, was allowed to select my own working hours, and he was always available online whenever I had questions. Even in the late evening or on weekends I could count on his help and advice. I would also like to express my sincere thanks to my major colaborators and inofficial co-advisors Thatchaphol Saranurak, Aaron Sidford, Yin Tat Lee, and Richard Peng. I would like to thank my other collaborators Aaron Bernstein, Joakim Blikstad, Maximilian Probst Gutenberg, Yang P. Liu, Sagnik Mukhopadhyay, Binghui Peng, Zhao Song, He Sun, Di Wang, and Omri Weinstein. I also thank Mikkel Thorup for inviting me to Copenhagen, the discussions we had, and the advice he has given while I was visiting. I thank Per Austrin for reading my thesis, helping with the Swedish summary, and for the enjoyable coffee and lunch breaks. I want to thank my fellow PhD students at KTH: Mohit, for repeatedly pushing me out of my comfort zone, e.g. to apply for the Google PhD Fellowship. Andreas, for lending me an ear whenever I felt like complaining about our landlords, and also for taking care of my apartment whenever I was traveling. Joseph, for introducing me to web fiction and the discus- sions we had. I also thank him, Kilian and Stephan for being available whenever I needed a break. Before coming to KTH, I studied at the Goethe University in Frankfurt and I want to thank faculty there for preparing me for my time as a PhD student. Thorsten Theobald, who was my Liaison Professor at the “Studienstiftung”, has offered me a lot of advice during my time at the Goethe University. He, Amin Coja-Oghlan and Rudolf Mester also offered a lot of advice near the end of my time at the Goethe University regarding how to proceed after graduation. I am also deeply grateful to Amin for allowing me to finish my Master’s Thesis while already working in Sweden as a PhD student. I thank Ronja D¨uffeland Hartwig Bosse for their support and advice related to teaching, and Hartwig’s advice on writing has been very helpful during my PhD. And finally, I thank my family and friends in Frankfurt. My friends always made time for me whenever I visited, even when the visits turned out to be unexpectedly iii iv frequent. (Again, thanks to Danupon for allowing me this freedom to frequently visit Frankfurt.) My parents who supported me all my life and who have taken great effort to provide me with the opportunity to focus on my education. At last, thank you, Stefanie, for brightening my life, no matter the distance between us. Contents Contents v I Thesis 1 1 Introduction 3 1.1 Dynamic Algorithms . 4 1.2 Optimization . 6 2 Publication List 11 3 Dynamic Linear Algebra 13 3.1 Unifying Matrix Data Structures . 13 3.2 Dynamic Matrix Inverse . 14 4 Dynamic Distances 21 4.1 New Dynamic Algorithms . 22 4.2 From Dynamic Linear Algebra to Dynamic Distances . 24 5 Convex and Combinatorial Optimization 27 5.1 Algorithmic Results . 28 5.2 From Dynamic Linear Algebra to Optimization . 31 II Included Papers 47 A Dynamic Matrix Inverse 49 A.1 Introduction .
Recommended publications
  • Stochastic Combinatorial Optimization with Controllable Risk Aversion Level
    Stochastic Combinatorial Optimization with Controllable Risk Aversion Level Anthony Man–Cho So Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Shatin, N. T., Hong Kong email: [email protected] Jiawei Zhang Department of Information, Operations, and Management Sciences, Stern School of Business, New York University, New York, NY 10012, USA email: [email protected] Yinyu Ye Department of Management Science and Engineering and, by courtesy, Electrical Engineering, Stanford University, Stanford, CA 94305, USA email: [email protected] Most of the recent work on 2–stage stochastic combinatorial optimization problems have focused on the min- imization of the expected cost or the worst–case cost of the solution. Those two objectives can be viewed as two extreme ways of handling risk. In this paper we propose to use an one–parameter family of functionals to interpolate between them. Although such a family has been used in the mathematical finance and stochastic programming literature before, its use in the context of approximation algorithms seems new. We show that under standard assumptions, a broad class of risk–adjusted 2–stage stochastic programs can be efficiently treated by the Sample Average Approximation (SAA) method. In particular, our result shows that it is computationally feasible to incorporate some degree of robustness even when the underlying distribution can only be accessed in a black–box fashion. We also show that when combined with suitable rounding procedures, our result yields new approximation algorithms for many risk–adjusted 2–stage stochastic combinatorial optimization problems under the black–box setting.
    [Show full text]
  • Opf)P in OPERATIONS RESEARCH
    APPLICATION OF MATHEMATICAL PROGRAMMING DISSERTATION SUBMITTED IN PARTIAL FULFILMENT OF THE REQUIREMENTS FOR THE AWARD OF THE DEGREE OF iWasiter of ^i)Uo«opf)p IN OPERATIONS RESEARCH By TEG ALAM Under the Supervision of DR. A. BARI DEPARTMENT OF STATISTICS & OPERATIONS-RESEARCH ALIGARH MUSLIM UNIVERSITY ALIGARH (INDIA) 2001 '^•^ P5.-32^^ ^, DS3200 DEPARTMENT OF STATISTICS & OPERATION RESEARCI ALIGARH MUSLIM UNIVERSITY (Dr. X ^ari ALIGARH - 202002 -INDIA Ph.D (AUg) ^hone: 0571-701251 (O) "^ 0571-700112 (R) Dated: OE(RTlTICA'VE Certified that the dissertation entitled "Application of Mathematical Programming" is carried out by Tcg Alaitl under my supervision. The work is sufficient for the requirement of degree of Master of Philosophy in Operations Research. ((Dr. A. (Bari) Supervisor Univ. Exchange: (0571) 700920 -23 Extns:419, 421,422, 441 Fax:+91-571-400528 This Dissertation entitled ''^Application of Mathematical programming''' is submitted to the Aligarh Muslim University, Aligarh, for the partial fiilfillment of the degree of M.Phil in OperationsResearch. Mathematical programming is concerned with the determination of a minimum or a maximum of a function of several variables, which are required to satisfy a number of constraints. Such solutions are sought in diverse fields; including engineering, Operations Research, Management Science, numerical analysis and economics etc. This mamuscript consists of five chapters. Chapter-1 deals with the brief history of Mathematical Programming. It also contains numerous applications of Mathematical Programming. Chapter-2 gives an introduction of Transportation problem and devoted to extensions and methods of solutions. A numerical example and Relation of Transportation problems with Network problems are also discussed. Chapter - 3 gives an introduction of Game theory.
    [Show full text]
  • Optima 82 Publishes the Obituary of Paul Tseng by His Friends Affected by the Proposed Change; Only Their Subtitles Would Change
    OPTIMA Mathematical Programming Society Newsletter 82 Steve Wright MPS-sponsored meetings: the International Conference on Stochas- MPS Chair’s Column tic Programming (ICSP) XII (Halifax, August 14–20), the Interna- tional Conference on Engineering Optimization (Lisbon, September 6–9), and the IMA Conference on Numerical Linear Algebra and March 16, 2010. You should recently have received a letter con- Optimization (Birmingham, September 13–15). cerning a possible change of name for MPS, to “Mathematical Op- timization Society”. This issue has been discussed in earnest since ISMP 2009, where it was raised at the Council and Business meet- ings. Some of you were kind enough to send me your views follow- Note from the Editors ing the mention in my column in Optima 80. Many (including me) believe that the term “optimization” is more widely recognized and The topic of this issue of Optima is Mechanism Design –aNobel better understood as an appellation for our field than the current prize winning theoretical field of economics. name, both among those working in the area and our colleagues We present the main article by Jay Sethuraman, which introduces in other disciplines. Others believe that the current name should the main concepts and existence results for some of the models be retained, as it has the important benefits of tradition, branding, arising in mechanism design theory. The discussion column by Garud and name recognition. To ensure archival continuity in the literature, Iyengar and Anuj Kumar address a specific example of such a model the main titles of our journals Mathematical Programming, Series A which can be solved by the means of optimization.
    [Show full text]
  • Arxiv:1411.2129V1 [Math.OC]
    INTERIOR-POINT ALGORITHMS FOR CONVEX OPTIMIZATION BASED ON PRIMAL-DUAL METRICS TOR MYKLEBUST, LEVENT TUNC¸EL Abstract. We propose and analyse primal-dual interior-point algorithms for convex opti- mization problems in conic form. The families of algorithms we analyse are so-called short- step algorithms and they match the current best iteration complexity bounds for primal-dual symmetric interior-point algorithm of Nesterov and Todd, for symmetric cone programming problems with given self-scaled barriers. Our results apply to any self-concordant barrier for any convex cone. We also prove that certain specializations of our algorithms to hyperbolic cone programming problems (which lie strictly between symmetric cone programming and gen- eral convex optimization problems in terms of generality) can take advantage of the favourable special structure of hyperbolic barriers. We make new connections to Riemannian geometry, integrals over operator spaces, Gaussian quadrature, and strengthen the connection of our algorithms to quasi-Newton updates and hence first-order methods in general. arXiv:1411.2129v1 [math.OC] 8 Nov 2014 Date: November 7, 2014. Key words and phrases. primal-dual interior-point methods, convex optimization, variable metric methods, local metric, self-concordant barriers, Hessian metric, polynomial-time complexity; AMS subject classification (MSC): 90C51, 90C22, 90C25, 90C60, 90C05, 65Y20, 52A41, 49M37, 90C30. Tor Myklebust: Department of Combinatorics and Optimization, Faculty of Mathematics, University of Wa- terloo, Waterloo, Ontario N2L 3G1, Canada (e-mail: [email protected]). Research of this author was supported in part by an NSERC Doctoral Scholarship, ONR Research Grant N00014-12-10049, and a Dis- covery Grant from NSERC.
    [Show full text]
  • A Primal-Dual Interior-Point Algorithm for Nonsymmetric Exponential-Cone Optimization
    A primal-dual interior-point algorithm for nonsymmetric exponential-cone optimization Joachim Dahl Erling D. Andersen May 27, 2019 Abstract A new primal-dual interior-point algorithm applicable to nonsymmetric conic optimization is proposed. It is a generalization of the famous algorithm suggested by Nesterov and Todd for the symmetric conic case, and uses primal-dual scalings for nonsymmetric cones proposed by Tun¸cel.We specialize Tun¸cel'sprimal-dual scalings for the important case of 3 dimensional exponential-cones, resulting in a practical algorithm with good numerical performance, on level with standard symmetric cone (e.g., quadratic cone) algorithms. A significant contribution of the paper is a novel higher-order search direction, similar in spirit to a Mehrotra corrector for symmetric cone algorithms. To a large extent, the efficiency of our proposed algorithm can be attributed to this new corrector. 1 Introduction In 1984 Karmarkar [11] presented an interior-point algorithm for linear optimization with polynomial complexity. This triggered the interior-point revolution which gave rise to a vast amount research on interior-point methods. A particularly important result was the analysis of so-called self-concordant barrier functions, which led to polynomial-time algorithms for linear optimization over a convex domain with self-concordant barrier, provided that the barrier function can be evaluated in polynomial time. This was proved by Nesterov and Nemirovski [18], and as a consequence convex optimization problems with such barriers can be solved efficently by interior-point methods, at least in theory. However, numerical studies for linear optimization quickly demonstrated that primal-dual interior- point methods were superior in practice, which led researchers to generalize the primal-dual algorithm to general smooth convex problems.
    [Show full text]
  • George B. Dantzig Papers SC0826
    http://oac.cdlib.org/findaid/ark:/13030/c8s75gwd No online items Guide to the George B. Dantzig Papers SC0826 Daniel Hartwig & Jenny Johnson Department of Special Collections and University Archives March 2012 Green Library 557 Escondido Mall Stanford 94305-6064 [email protected] URL: http://library.stanford.edu/spc Guide to the George B. Dantzig SC0826 1 Papers SC0826 Language of Material: English Contributing Institution: Department of Special Collections and University Archives Title: George B. Dantzig papers creator: Dantzig, George Bernard, 1914-2005 Identifier/Call Number: SC0826 Physical Description: 91 Linear Feet Date (inclusive): 1937-1999 Special Collections and University Archives materials are stored offsite and must be paged 36-48 hours in advance. For more information on paging collections, see the department's website: http://library.stanford.edu/depts/spc/spc.html. Information about Access The materials are open for research use. Audio-visual materials are not available in original format, and must be reformatted to a digital use copy. Ownership & Copyright All requests to reproduce, publish, quote from, or otherwise use collection materials must be submitted in writing to the Head of Special Collections and University Archives, Stanford University Libraries, Stanford, California 94305-6064. Consent is given on behalf of Special Collections as the owner of the physical items and is not intended to include or imply permission from the copyright owner. Such permission must be obtained from the copyright owner, heir(s) or assigns. See: http://library.stanford.edu/depts/spc/pubserv/permissions.html. Restrictions also apply to digital representations of the original materials. Use of digital files is restricted to research and educational purposes.
    [Show full text]
  • Semidefinite Relaxation of Quadratic Optimization Problems
    APPEARED IN IEEE SP MAGAZINE, SPECIAL ISSUE ON “CONVEX OPT. FOR SP”, MAY 2010 1 Semidefinite Relaxation of Quadratic Optimization Problems Zhi-Quan Luo, Wing-Kin Ma, Anthony Man-Cho So, Yinyu Ye, and Shuzhong Zhang 2 I. INTRODUCTION In recent years, the semidefinite relaxation (SDR) technique 1.5 has been at the center of some of the very exciting develop- 1 ments in the area of signal processing and communications, and it has shown great significance and relevance on a va- 0.5 riety of applications. Roughly speaking, SDR is a powerful, 2 0 x computationally efficient approximation technique for a host of very difficult optimization problems. In particular, it can be −0.5 applied to many nonconvex quadratically constrained quadratic −1 programs (QCQPs) in an almost mechanical fashion. These include the following problems: −1.5 T min x Cx −2 Rn −2 −1.5 −1 −0.5 0 0.5 1 1.5 2 x∈ x1 T s.t. x Fix gi, i =1, . , p, (1) T ≥ R2 x Hix = li, i =1, . , q, Fig. 1. A nonconvex QCQP in . Colored lines: contour of the objective function; gray area: the feasible set; black lines: boundary of each constraint. where the given matrices C, F1,..., Fp, H1,..., Hq are as- sumed to be general real symmetric matrices, possibly in- definite. The class of nonconvex QCQPs (1) captures many accuracy of no worse than 0.8756 for the Maximum Cut problems that are of interest to the signal processing and com- problem (the BQP with some conditions on C). In other munications community.
    [Show full text]
  • Low-Rank Semidefinite Programming: Theory And
    Full text available at: http://dx.doi.org/10.1561/2400000009 Low-Rank Semidefinite Programming: Theory and Applications Alex Lemon Stanford University [email protected] Anthony Man-Cho The Chinese University of Hong Kong [email protected] Yinyu Ye Stanford University [email protected] Boston — Delft Full text available at: http://dx.doi.org/10.1561/2400000009 R Foundations and Trends in Optimization Published, sold and distributed by: now Publishers Inc. PO Box 1024 Hanover, MA 02339 United States Tel. +1-781-985-4510 www.nowpublishers.com [email protected] Outside North America: now Publishers Inc. PO Box 179 2600 AD Delft The Netherlands Tel. +31-6-51115274 The preferred citation for this publication is A. Lemon, A. M.-C. So, Y. Ye. Low-Rank Semidefinite Programming: Theory and Applications. Foundations and Trends R in Optimization, vol. 2, no. 1-2, pp. 1–156, 2015. R This Foundations and Trends issue was typeset in LATEX using a class file designed by Neal Parikh. Printed on acid-free paper. ISBN: 978-1-68083-137-5 c 2016 A. Lemon, A. M.-C. So, Y. Ye All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, mechanical, photocopying, recording or otherwise, without prior written permission of the publishers. Photocopying. In the USA: This journal is registered at the Copyright Clearance Cen- ter, Inc., 222 Rosewood Drive, Danvers, MA 01923. Authorization to photocopy items for internal or personal use, or the internal or personal use of specific clients, is granted by now Publishers Inc for users registered with the Copyright Clearance Center (CCC).
    [Show full text]
  • The 2010 Noticesindex
    The 2010 Notices Index January issue: 1–200 May issue: 593–696 October issue: 1073–1240 February issue: 201–328 June/July issue: 697–816 November issue: 1241–1384 March issue: 329–456 August issue: 817–936 December issue: 1385–1552 April issue: 457–592 September issue: 937–1072 2010 Award for an Exemplary Program or Achievement in Index Section Page Number a Mathematics Department, 650 2010 Conant Prize, 515 The AMS 1532 2010 E. H. Moore Prize, 524 Announcements 1533 2010 Mathematics Programs that Make a Difference, 650 Articles 1533 2010 Morgan Prize, 517 Authors 1534 2010 Robbins Prize, 526 Deaths of Members of the Society 1536 2010 Steele Prizes, 510 Grants, Fellowships, Opportunities 1538 2010 Veblen Prize, 521 Letters to the Editor 1540 2010 Wiener Prize, 519 Meetings Information 1540 2010–2011 AMS Centennial Fellowship Awarded, 758 New Publications Offered by the AMS 1540 2011 AMS Election—Nominations by Petition, 1027 Obituaries 1540 AMS Announces Congressional Fellow, 765 Opinion 1540 AMS Announces Mass Media Fellowship Award, 766 Prizes and Awards 1541 AMS-AAAS Mass Media Summer Fellowships, 1320 Prizewinners 1543 AMS Centennial Fellowships, Invitation for Applications Reference and Book List 1547 for Awards, 1320 Reviews 1547 AMS Congressional Fellowship, 1323 Surveys 1548 AMS Department Chairs Workshop, 1323 Tables of Contents 1548 AMS Email Support for Frequently Asked Questions, 268 AMS Endorses Postdoc Date Agreement, 1486 AMS Homework Software Survey, 753 About the Cover AMS Holds Workshop for Department Chairs, 765 57,
    [Show full text]
  • Prof. Yinyu Ye
    静园 号院 科研讲座 2 0 1 9 年第03号 Distributionally Robust Optimization Driven by Stochastic and Online Data Prof. Yinyu Ye 邓小铁 教授 2019年4月10日 星期三 15:10-16:10 静园五院107教室 Abstract We present decision/optimization models/problems driven by uncertain and online data, and show how analytical models and computational algorithms can be used to achieve solution efficiency and near optimality. First, we describe the so-called Distributionally or Likelihood Robust optimization (DRO) models and their algorithms in dealing stochastic decision problems when the exact uncertainty distribution is unknown but certain statistical moments and/or sample distributions can be estimated. Secondly, when decisions are made in presence of high dimensional stochastic data, handling joint distribution of correlated random variables can present a formidable task, both in terms of sampling and estimation as well as algorithmic complexity. A common heuristic is to estimate only marginal distributions and substitute joint distribution by independent (product) distribution. Here, we study possible loss incurred on ignoring correlations through the DRO approach, and quantify that loss as Price of Correlations (POC). Thirdly, we describe an online combinatorial auction problem using online linear programming technologies. We discuss near-optimal algorithms for solving this surprisingly general class of online problems under the assumption of random order of arrivals and some conditions on the data and size of the problem. Biography Yinyu Ye is currently the K.T. Li Chair Professor of Engineering at Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering, Stanford University. He is also the Director of the MS&E Industrial Affiliates Program.
    [Show full text]
  • 1 Plenary Talks of Mostly OM 2019 Workshop Dimitris Bertsimas
    Plenary Talks of Mostly OM 2019 Workshop Dimitris Bertsimas Massachusetts Institute of Technology Title: The Voice of Optimization Abstract: We present a new way to see optimization problems. Using machine learning techniques we are able to predict the strategy behind the optimal solution in any continuous and mixed-integer convex optimization problem as a function of its key parameters. The benefits of our approach are interpretability and speed. We use interpretable machine learning algorithms such as optimal classification trees (OCTs) to gain insights on the relationship between the problem parameters and the optimal solution. In this way, optimization is no longer a black-box and we can understand it. In addition, once we train the predictor, we can solve optimization problems at very high speed. This aspect is also relevant for non interpretable machine learning methods such as neural networks (NNs) since they can be evaluated very efficiently after the training phase. We show on several realistic examples that the accuracy behind our approach is in the 90%-100% range, while even when the predictions are not correct, the degree of suboptimality or infeasibility is very low. We also benchmark the computation time beating state of the art solvers by multiple orders of magnitude. Therefore, our method provides on the one hand a novel insightful understanding of the optimal strategies to solve a broad class of continuous and mixed-integer optimization problems and on the other hand a powerful computational tool to solve online optimization at very high speed. (joint work with Bartolomeo Stellato, MIT). About the speaker: Dimitris Bertsimas is currently the Boeing Professor of Operations Research and the co- director of the Operations Research Center and faculty director of the Master of Business analytics at MIT.
    [Show full text]
  • Yinyu Ye Received the B.S. Degree in System Engineering from the Huazhong University of Science and Technology, Wuhan, China, and the M.S
    Yinyu Ye received the B.S. degree in System Engineering from the Huazhong University of Science and Technology, Wuhan, China, and the M.S. and Ph.D. degrees in Management Science & Engineering from Stanford University, Stanford. Currently, he is a full Professor of Management Science and Engineering and Institute of Computational and Mathematical Engineering and the Director of the MS&E Industrial Affiliates Program, Stanford University. His current research interests include Continuous and Discrete Optimization, Mathematical Programming, Algorithm Design and Analysis, Computational Game/Market Equilibrium, Metric Distance Geometry, Graph Realization, Dynamic Resource Allocation, and Stochastic and Robust Decision Making, etc. Highlights of his academic research include: J. Carlsson (supervised by Ye), the 2013 INFORMS Computing Society Prize, for the best English language papers dealing with the Operations Research/Computer Science interface I. Post, the 2013 Second Prize of INFORMS Nicholson Student Paper Competition for the joint paper with Ye “The simplex method is strongly polynomial for Markov decision processes.” 2012 ISMP Tseng Lectureship Prize (Inaugural Recipient) for outstanding contributions in the area of continuous optimization Recipient of the 2009 John von Neumann Theory Prize for fundamental sustained contributions to theory in Operations Research and the Management Sciences Recipient of the 2009 IBM Faculty Award. Elected Vice Chair of the SIAM Activity Group on Optimization (SIAG/OPT), 2008. E. Delage, First
    [Show full text]