Affine-Scaling Method, 202–204, 206, 210 Poor Performance Of

Total Page:16

File Type:pdf, Size:1020Kb

Affine-Scaling Method, 202–204, 206, 210 Poor Performance Of Index affine-scaling method, 202–204, 206, 210 blocking variable, 48 poor performance of, 203 breakpoint, 163 steplength choice, 203 approximation problems, 218–227 canonical form, 8, 16, 45, 117, 120, 123, 1-norm, 221–224 129, 130, 134, 138, 139, 150, 2-norm. See least-squares problems 151, 156, 195, 197 ∞-norm. See Chebyshev approximation dual of, 109 for linear equalities, 218–223 for problems with bound constraints, with hard constraints, 223–224, 227 130 arc, 12–14, 144 centering parameter, 206, 214 central path, 205 back substitution. See triangular substitution Chebyshev approximation, 219–221, 223 basic feasible solution, 118 Cholesky factorization, 209, 212 as iterates of revised simplex method, classification, 11–12, 230–235 123 labeled points, 230 association with basis matrix, 120–121 testing set, 235 association with vertex, 121–123 training set, 235 degenerate, 120 tuning set, 235 existence, 119–120 classifier, 230 initial, 129, 134–139 linear, 230 basic solution, 118 nonlinear, 233, 234 basis, 117, 126, 152 complementarity, 101–102, 178, 237 for problems with bound constraints, almost-, 178, 179 130 existence of strictly complementary initial, 125, 134–137 primal-dual solution, 114–115 optimal, 129, 152–156 strict, 101 basis matrix, 52, 118, 120, 123, 139, 140, complementary slackness. 154 See complementarity for network linear program, 149 concave function, 161, 244 LU factorization of, 139–142 constraints, 1 big M method, 110–112 active, 14 behavior on infeasible problems, 112 bounds, 74–75 motivation, 110–111 equality, 72–87 proof of effectiveness, 111 inactive, 14 specification, 111–112 inequality, 72–87 bimatrix game, 185 convex combination, 239, 241 Bland’s rule. See smallest-subscript rule convex function, 170, 242–244, 248 261 Copyright ©2007 by the Society for Industrial and Applied Mathematics This electronic version is for personal use and may not be duplicated or distributed. 262 Index linear underestimation, 170, 250 feasible region, 3, 46, 51, 53, 169, 171, strict, 173 182 convex hull, 56, 113, 241 feasible set. See feasible region convex set, 170, 173, 239–241 feature space, 233 cost vector, 7, 151, 154 fickle variable, 69 parametrized, 159 finite termination of simplex method, cross-validation, 235 65–72 cycling, 66–72, 104 floating-point number, 127 floating-point operations, 39 Dantzig, George B., 7, 16 forward substitution. See triangular degenerate linear program, 66, 67 substitution diet problem, 8, 96 Frank–Wolfe theorem, 172–173 divergence, 144 fundamental theorem of algebra, 113, 225 domain, 242 dual linear program Gaussian elimination, 39 of canonical form, 196 general form of linear program, 75 of standard form, 94, 102 dual of, 107–109 tableau representation, 94 global solution, 170, 171, 173, 182 dual linear system, 89 Gordan theorem, 113 dual pair of linear programs. See primal- gradient, 172, 242, 244, 247 dual pair of linear programs graphical solution of linear programs, 2–6 dual simplex method, 89, 102–106, 164, 220–223 Hessian, 171, 172, 214, 224, 242, 244, motivation for, 102 245 relationship to simplex method applied Huber estimation to the dual, 105–106 formulation as quadratic program, specification of, 103–104 229–230 duality, 89–115 motivation and definition, 227–228 for quadratic programs, 176–177 optimality conditions, 228–229 practical example, 96–98 strong, 98–99 infeasible linear program, 14, 62, 98, 225 applications of, 112–115, 167 infeasible point, 3 for quadratic programming, 176–177 integer programming, 7, 16 weak, 97–98, 101, 111, 176 interior-point methods duality gap, 197 basic properties of primal-dual duality measure, 197, 205, 213 methods, 196–197 comparison with simplex method, 212 ellipsoid method, 15 introduction, 15, 195–197 entering variable, 47 path-following. See path-following epigraph, 217, 242–243 methods equilibrium pair. See Nash equilibrium relationship to Newton’s method, 202 expected loss, 185 Jacobian, 199, 202, 249 Farkas lemma, 112–113 Jordan exchange, 17–23, 26, 27, 46–48, feasible point, 3 53, 83, 179, 221. See also pivot dual, 98, 103 block, 31 Copyright ©2007 by the Society for Industrial and Applied Mathematics This electronic version is for personal use and may not be duplicated or distributed. Index 263 blocking of, 25, 33, 36, 38 relationship to quadratic programming, interpretation on dual tableaus, 89–91 174, 177, 180, 182 linear dependence, 23, 24, 138, 221 Karmarkar, Narendra, 15, 214 of functions, 24 Karush–Kuhn–Tucker (KKT) conditions, linear equations, 32–39 100–101, 129, 192, 195 inconsistent, 218, 227 as constrained nonlinear equations, overdetermined, 220, 221, 223, 224, 201–202 226, 227, 230 for canonical form, 195, 196 solution using Gaussian elimination, for linear programs in general form, 39–41 108 solution using Jordan exchanges, for quadratic programs, 173–175, 181, 32–39 184–185, 212, 229, 233 flop count, 39, 44 statement of, 100–101 with infinitely many solutions, 37 kernel, 234 linear independence, 17, 23–27, 51, 52, Gaussian, 234–235 118, 119 linear, 234 of functions, 24 polynomial, 234 of rows/columns in a matrix, 92 KKT (Karush–Kuhn–Tucker) conditions. local solution, 169, 170, 173, 184 See Karush–Kuhn–Tucker LU decomposition. See LU factorization conditions LU factorization, 17, 39–44, 127, 138, 197 knapsack problem, 130 complete pivoting, 43 flop count, 44 partial pivoting, 42–44 Lagrange multipliers, 173, 176, 195, 229, updating, 139–142 232 Lagrangian, 176 matching pennies game, 188–191, 193 LCP. See linear complementarity matrix problem (LCP) addition, 239 least-squares problems, 211, 224–229 basis. See basis matrix normal equations, 225 diagonal, 196, 208, 238 leaving variable, 48, 156 eigenvalues, 246 Lemke’s method, 178–185 identity, 238 application to quadratic programs, indefinite, 208, 214, 245 182–185 inverse, 27–31, 238 outline of, 178 invertible, 51, 152, 225 Phase I, 178–179, 188 loss, 185–188, 191 Phase II, 179, 188 lower triangular, 39, 209, 239 termination of, 182 unit, 40 Lemke–Howson method, 187 multiplication, 239 level set, 243 nonsingular, 27 linear combination, 240 permutation, 39–41, 139, 209, 238 linear complementarity problem (LCP), poorly conditioned, 226 169, 177–178 positive definite, 245, 246 definition, 177 positive semidefinite, 172, 174, 176, infeasible, 182 224, 245, 246 Copyright ©2007 by the Society for Industrial and Applied Mathematics This electronic version is for personal use and may not be duplicated or distributed. 264 Index representation in MATLAB,251–252 p, 219, 247 singular, 27, 31 ∞, 219, 247 skew-symmetric, 178 dual, 231, 248 sparse, 208, 209, 238 equivalence of, 248 symmetric, 172, 244 totally unimodular, 150 objective function, 1, 154, 241 transpose, 238 contours, 4, 46, 159 upper triangular, 39 optimal face, 58. See also solution set Mehrotra predictor-corrector method. order notation See path-following methods O(·), 248 minimax problems, 217–218 o(·), 248–250 discrete, 218 minimum principle, 170, 173, 184 parametric programming, 151, 158–168 mixed strategy, 185 path-following methods mixed-integer linear programming, 150 MATLAB implementation of, 209 monotone, 247 choice of centering parameter, 207, Nash equilibrium, 186, 192 210 computation of, 186–187 choice of starting point, 211 Nash, John, 169, 186 choice of steplength, 207, 211 network linear program, 143–150 corrector direction, 210, 211 assignment problem, 149 for quadratic programming, 212–214 circulation problem, 147 linear algebra issues, 208–209, 211, general properties, 143–144 214 max-flow, 146–147 normal-equations form, 208 minimum-cost, 12–14, 144–146, 149 long-step, 206–207 node-arc incidence matrix, 145 specification of algorithm, 206–207 shortest-path, 145–146 motivation, 204–205 transportation problem, 147–149 practical enhancements, 209–211 network simplex method, 149–150 relationship to Newton’s method, Newton’s method, 197–201 205–206 for scalar nonlinear equations, 197–198 permutation, 29 for systems of nonlinear equations, Phase I of simplex method, 60–65, 179, 197–201 223 quadratic convergence of, 199 after addition of constraints, 157 with line search, 201 description of, 60–65 node, 12, 144 dual simplex as possible alternative, demand, 13, 145 102, 104 sink, 145 for canonical form, 134–138 supply, 13 for problems with parametrized con- nondegenerate linear program, 65, 66 straints, 167 nonlinear programming, 169–172 for solution of linear inequalities, 64 nonnegative orthant, 171, 251 purpose of, 53 norm, 218, 247–248 Phase II of simplex method, 53–60, 62, 1, 221, 247 71, 103, 135–137 2, 231, 247 piecewise-linear function, 161, 163, 217 Copyright ©2007 by the Society for Industrial and Applied Mathematics This electronic version is for personal use and may not be duplicated or distributed. Index 265 pivot, 18, 20, 46, 47. See also Jordan exchange definition, 49 block, 31 for dual simplex method, 103, 104, degenerate, 62, 64, 69 157, 165 nondegenerate, 67 in Lemke’s method, 179, 180, 188 submatrix, 31 robust implementation of, 143 pivot selection, 15. See also pricing ties in, 66, 68 Devex, 142, 143 reduced costs, 49, 68, 124, 132, 142, 152, for dual simplex, 103 153, 160 for Lemke’s algorithm, 179 regression problems. See approximation steepest-edge, 142 problems polynomial complexity, 207 residual vector, 219, 221, 227 predictor-corrector method. outliers, 227 See path-following methods resource allocation, 10–11 preprocessing, 212 revised simplex method, 52, 117, 123–143, pricing, 47, 49, 53, 68, 117 156 partial, 142 choice of initial basis, 125 primal-dual methods. See interior-point for
Recommended publications
  • Using Ant System Optimization Technique for Approximate Solution to Multi- Objective Programming Problems
    International Journal of Scientific & Engineering Research, Volume 4, Issue 9, September-2013 ISSN 2229-5518 1701 Using Ant System Optimization Technique for Approximate Solution to Multi- objective Programming Problems M. El-sayed Wahed and Amal A. Abou-Elkayer Department of Computer Science, Faculty of Computers and Information Suez Canal University (EGYPT) E-mail: [email protected], Abstract In this paper we use the ant system optimization metaheuristic to find approximate solution to the Multi-objective Linear Programming Problems (MLPP), the advantageous and disadvantageous of the suggested method also discussed focusing on the parallel computation and real time optimization, it's worth to mention here that the suggested method doesn't require any artificial variables the slack and surplus variables are enough, a test example is given at the end to show how the method works. Keywords: Multi-objective Linear. Programming. problems, Ant System Optimization Problems Introduction The Multi-objective Linear. Programming. problems is an optimization method which can be used to find solutions to problems where the Multi-objective function and constraints are linear functions of the decision variables, the intersection of constraints result in a polyhedronIJSER which represents the region that contain all feasible solutions, the constraints equation in the MLPP may be in the form of equalities or inequalities, and the inequalities can be changed to equalities by using slack or surplus variables, one referred to Rao [1], Philips[2], for good start and introduction to the MLPP. The LP type of optimization problems were first recognized in the 30's by economists while developing methods for the optimal allocations of resources, the main progress came from George B.
    [Show full text]
  • An Annotated Bibliography of Network Interior Point Methods
    AN ANNOTATED BIBLIOGRAPHY OF NETWORK INTERIOR POINT METHODS MAURICIO G.C. RESENDE AND GERALDO VEIGA Abstract. This paper presents an annotated bibliography on interior point methods for solving network flow problems. We consider single and multi- commodity network flow problems, as well as preconditioners used in imple- mentations of conjugate gradient methods for solving the normal systems of equations that arise in interior network flow algorithms. Applications in elec- trical engineering and miscellaneous papers complete the bibliography. The collection includes papers published in journals, books, Ph.D. dissertations, and unpublished technical reports. 1. Introduction Many problems arising in transportation, communications, and manufacturing can be modeled as network flow problems (Ahuja et al. (1993)). In these problems, one seeks an optimal way to move flow, such as overnight mail, information, buses, or electrical currents, on a network, such as a postal network, a computer network, a transportation grid, or a power grid. Among these optimization problems, many are special classes of linear programming problems, with combinatorial properties that enable development of efficient solution techniques. The focus of this annotated bibliography is on recent computational approaches, based on interior point methods, for solving large scale network flow problems. In the last two decades, many other approaches have been developed. A history of computational approaches up to 1977 is summarized in Bradley et al. (1977). Several computational studies established the fact that specialized network simplex algorithms are orders of magnitude faster than the best linear programming codes of that time. (See, e.g. the studies in Glover and Klingman (1981); Grigoriadis (1986); Kennington and Helgason (1980) and Mulvey (1978)).
    [Show full text]
  • Matroidal Subdivisions, Dressians and Tropical Grassmannians
    Matroidal subdivisions, Dressians and tropical Grassmannians vorgelegt von Diplom-Mathematiker Benjamin Frederik Schröter geboren in Frankfurt am Main Von der Fakultät II – Mathematik und Naturwissenschaften der Technischen Universität Berlin zur Erlangung des akademischen Grades Doktor der Naturwissenschaften – Dr. rer. nat. – genehmigte Dissertation Promotionsausschuss: Vorsitzender: Prof. Dr. Wilhelm Stannat Gutachter: Prof. Dr. Michael Joswig Prof. Dr. Hannah Markwig Senior Lecturer Ph.D. Alex Fink Tag der wissenschaftlichen Aussprache: 17. November 2017 Berlin 2018 Zusammenfassung In dieser Arbeit untersuchen wir verschiedene Aspekte von tropischen linearen Räumen und deren Modulräumen, den tropischen Grassmannschen und Dressschen. Tropische lineare Räume sind dual zu Matroidunterteilungen. Motiviert durch das Konzept der Splits, dem einfachsten Fall einer polytopalen Unterteilung, wird eine neue Klasse von Matroiden eingeführt, die mit Techniken der polyedrischen Geometrie untersucht werden kann. Diese Klasse ist sehr groß, da sie alle Paving-Matroide und weitere Matroide enthält. Die strukturellen Eigenschaften von Split-Matroiden können genutzt werden, um neue Ergebnisse in der tropischen Geometrie zu erzielen. Vor allem verwenden wir diese, um Strahlen der tropischen Grassmannschen zu konstruieren und die Dimension der Dressschen zu bestimmen. Dazu wird die Beziehung zwischen der Realisierbarkeit von Matroiden und der von tropischen linearen Räumen weiter entwickelt. Die Strahlen einer Dressschen entsprechen den Facetten des Sekundärpolytops eines Hypersimplexes. Eine besondere Klasse von Facetten bildet die Verallgemeinerung von Splits, die wir Multi-Splits nennen und die Herrmann ursprünglich als k-Splits bezeichnet hat. Wir geben eine explizite kombinatorische Beschreibung aller Multi-Splits eines Hypersimplexes. Diese korrespondieren mit Nested-Matroiden. Über die tropische Stiefelabbildung erhalten wir eine Beschreibung aller Multi-Splits für Produkte von Simplexen.
    [Show full text]
  • An Affine-Scaling Pivot Algorithm for Linear Programming
    AN AFFINE-SCALING PIVOT ALGORITHM FOR LINEAR PROGRAMMING PING-QI PAN¤ Abstract. We proposed a pivot algorithm using the affine-scaling technique. It produces a sequence of interior points as well as a sequence of vertices, until reaching an optimal vertex. We report preliminary but favorable computational results obtained with a dense implementation of it on a set of standard Netlib test problems. Key words. pivot, interior point, affine-scaling, orthogonal factorization AMS subject classifications. 65K05, 90C05 1. Introduction. Consider the linear programming (LP) problem in the stan- dard form min cT x (1.1) subject to Ax = b; x ¸ 0; where c 2 Rn; b 2 Rm , A 2 Rm£n(m < n). Assume that rank(A) = m and c 62 range(AT ). So the trivial case is eliminated where the objective value cT x is constant over the feasible region of (1.1). The assumption rank(A) = m is not substantial either to the proposed algorithm,and will be dropped later. As a LP problem solver, the simplex algorithm might be one of the most famous and widely used mathematical tools in the world. Its philosophy is to move on the underlying polyhedron, from a vertex to adjacent vertex, along edges until an optimal vertex is reached. Since it was founded by G.B.Dantzig [8] in 1947, the simplex algorithm has been very successful in practice, and occupied a dominate position in this area, despite its infiniteness in case of degeneracy. However, it turned out that the simplex algorithm may require an exponential amount of time to solve LP problems in the worst-case [29].
    [Show full text]
  • On Some Polynomial-Time Algorithms for Solving Linear Programming Problems
    Available online a t www.pelagiaresearchlibrary.com Pelagia Research Library Advances in Applied Science Research, 2012, 3 (5):3367-3373 ISSN: 0976-8610 CODEN (USA): AASRFC On Some Polynomial-time Algorithms for Solving Linear Programming Problems B. O. Adejo and H. S. Adaji Department of Mathematical Sciences, Kogi State University, Anyigba _____________________________________________________________________________________________ ABSTRACT In this article we survey some Polynomial-time Algorithms for Solving Linear Programming Problems namely: the ellipsoid method, Karmarkar’s algorithm and the affine scaling algorithm. Finally, we considered a test problem which we solved with the methods where applicable and conclusions drawn from the results so obtained. __________________________________________________________________________________________ INTRODUCTION An algorithm A is said to be a polynomial-time algorithm for a problem P, if the number of steps (i. e., iterations) required to solve P on applying A is bounded by a polynomial function O(m n,, L) of dimension and input length of the problem. The Ellipsoid method is a specific algorithm developed by soviet mathematicians: Shor [5], Yudin and Nemirovskii [7]. Khachian [4] adapted the ellipsoid method to derive the first polynomial-time algorithm for linear programming. Although the algorithm is theoretically better than the simplex algorithm, which has an exponential running time in the worst case, it is very slow practically and not competitive with the simplex method. Nevertheless, it is a very important theoretical tool for developing polynomial-time algorithms for a large class of convex optimization problems. Narendra K. Karmarkar is an Indian mathematician; renowned for developing the Karmarkar’s algorithm. He is listed as an ISI highly cited researcher.
    [Show full text]
  • A Truncated SQP Algorithm for Large Scale Nonlinear Programming Problems
    ' : .v ' : V. • ' '^!V'> r" '%': ' W ' ::r'' NISTIR 4900 Computing and Applied Mathematics Laboratory A Truncated SQP Algorithm for Large Scale Nonlinear Programming Problems P.T. Boggs, J.W. Tolle, A.J. Kearsley August 1992 Technology Administration U.S. DEPARTMENT OF COMMERCE National Institute of Standards and Technology Gaithersburg, MD 20899 “QC— 100 .U56 4900 1992 Jc. NISTIR 4900 A Truncated SQP Algorithm for Large Scale Nonlinear Programming Problems P. T. Boggs J. W. Tone A. J. Kearsley U.S. DEPARTMENT OF COMMERCE Technology Administration National Institute of Standards and Technology Computing and Applied Mathematics Laboratory Applied and Computational Mathematics Division Gaithersburg, MD 20899 August 1992 U.S. DEPARTMENT OF COMMERCE Barbara Hackman Franklin, Secretary TECHNOLOGY ADMINISTRATION Robert M. White, Under Secretary for Technology NATIONAL INSTITUTE OF STANDARDS AND TECHNOLOGY John W. Lyons, Director si.- MT-- ’•,•• /'k /-.<• -Vj* #' ^ ( "> M !f>' J ' • ''f,i sgsHv.w,^. ^ , .!n'><^.'". tr 'V 10 i^ailCin'i.-'J/ > ' it i>1l • •»' W " %'. ' ' -fft ?" . , - '>! .-' .S A Truncated SQP Algorithm for Large Scale Nonlinear Programming Problems * Paul T. Boggs ^ Jon W. ToUe ^ Anthony J. Kearsley § August 4, 1992 Abstract We consider the inequality constrained nonlinear programming problem and an SQP algorithm for its solution. We are primarily concerned with two aspects of the general procedure, namely, the approximate solution of the quadratic program, and the need for an appropriate merit function. We first describe an (iterative) interior-point method for the quadratic programming subproblem that, no matter when it its terminated, yields a descent direction for a suggested new merit function. An algorithm based on ideas from trust-region and truncated Newton methods, is suggested and some of our preliminary numerical results are discussed.
    [Show full text]
  • A Bound Strengthening Method for Optimal Transmission Switching in Power Systems
    1 A Bound Strengthening Method for Optimal Transmission Switching in Power Systems Salar Fattahi∗, Javad Lavaei∗;+, and Alper Atamturk¨ ∗ ∗ Industrial Engineering and Operations Research, University of California, Berkeley + Tsinghua-Berkeley Shenzhen Institute, University of California, Berkeley. Abstract—This paper studies the optimal transmission switch- transmission switches) in the system. Furthermore, The Energy ing (OTS) problem for power systems, where certain lines are Policy Act of 2005 explicitly addresses the “difficulties of fixed (uncontrollable) and the remaining ones are controllable siting major new transmission facilities” and calls for the via on/off switches. The goal is to identify a topology of the power grid that minimizes the cost of the system operation while utilization of better transmission technologies [2]. satisfying the physical and operational constraints. Most of the Unlike in the classical network flows, removing a line from existing methods for the problem are based on first converting a power network may improve the efficiency of the network the OTS into a mixed-integer linear program (MILP) or mixed- due to physical laws. This phenomenon has been observed integer quadratic program (MIQP), and then iteratively solving and harnessed to improve the power system performance by a series of its convex relaxations. The performance of these methods depends heavily on the strength of the MILP or MIQP many authors. The notion of optimally switching the lines of formulations. In this paper, it is shown that finding the strongest a transmission network was introduced by O’Neill et al. [3]. variable upper and lower bounds to be used in an MILP or Later on, it has been shown in a series of papers that the MIQP formulation of the OTS based on the big-M or McCormick incorporation of controllable transmission switches in a grid inequalities is NP-hard.
    [Show full text]
  • An Affine-Scaling Interior-Point Method for Continuous Knapsack Constraints with Application to Support Vector Machines∗
    SIAM J. OPTIM. c 2011 Society for Industrial and Applied Mathematics Vol. 21, No. 1, pp. 361–390 AN AFFINE-SCALING INTERIOR-POINT METHOD FOR CONTINUOUS KNAPSACK CONSTRAINTS WITH APPLICATION TO SUPPORT VECTOR MACHINES∗ MARIA D. GONZALEZ-LIMA†, WILLIAM W. HAGER‡ , AND HONGCHAO ZHANG§ Abstract. An affine-scaling algorithm (ASL) for optimization problems with a single linear equality constraint and box restrictions is developed. The algorithm has the property that each iterate lies in the relative interior of the feasible set. The search direction is obtained by approximating the Hessian of the objective function in Newton’s method by a multiple of the identity matrix. The algorithm is particularly well suited for optimization problems where the Hessian of the objective function is a large, dense, and possibly ill-conditioned matrix. Global convergence to a stationary point is established for a nonmonotone line search. When the objective function is strongly convex, ASL converges R-linearly to the global optimum provided the constraint multiplier is unique and a nondegeneracy condition holds. A specific implementation of the algorithm is developed in which the Hessian approximation is given by the cyclic Barzilai-Borwein (CBB) formula. The algorithm is evaluated numerically using support vector machine test problems. Key words. interior-point, affine-scaling, cyclic Barzilai-Borwein methods, global convergence, linear convergence, support vector machines AMS subject classifications. 90C06, 90C26, 65Y20 DOI. 10.1137/090766255 1. Introduction. In this paper we develop an interior-point algorithm for a box-constrained optimization problem with a linear equality constraint: (1.1) min f(x) subject to x ∈ Rn, l ≤ x ≤ u, aTx = b.
    [Show full text]
  • Interior Point Method
    LECTURE 6: INTERIOR POINT METHOD 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm Motivation • Simplex method works well in general, but suffers from exponential-time computational complexity. • Klee-Minty example shows simplex method may have to visit every vertex to reach the optimal one. • Total complexity of an iterative algorithm = # of iterations x # of operations in each iteration • Simplex method - Simple operations: Only check adjacent extreme points - May take many iterations: Klee-Minty example Question: any fix? • Complexity of the simplex method Worst case performance of the simplex method Klee-Minty Example: •Victor Klee, George J. Minty, “How good is the simplex algorithm?’’ in (O. Shisha edited) Inequalities, Vol. III (1972), pp. 159-175. Klee-Minty Example Klee-Minty Example Karmarkar’s (interior point) approach • Basic idea: approach optimal solutions from the interior of the feasible domain • Take more complicated operations in each iteration to find a better moving direction • Require much fewer iterations General scheme of an interior point method • An iterative method that moves in the interior of the feasible domain Interior movement (iteration) • Given a current interior feasible solution , we have > 0 An interior movement has a general format Key knowledge • 1. Who is in the interior? - Initial solution • 2. How do we know a current solution is optimal? - Optimality condition • 3. How to move to a new solution? - Which direction to move? (good feasible direction) - How far to go? (step-length) Q1 - Who is in the interior? • Standard for LP • Who is at the vertex? • Who is on the edge? • Who is on the boundary? • Who is in the interior? What have learned before Who is in the interior? • Two criteria for a point x to be an interior feasible solution: 1.
    [Show full text]
  • A Scenario-Based Branch-And-Bound Approach for MES Scheduling in Urban Buildings Mainak Dan, Student Member, Seshadhri Srinivasan, Sr
    1 A Scenario-based Branch-and-Bound Approach for MES Scheduling in Urban Buildings Mainak Dan, Student Member, Seshadhri Srinivasan, Sr. Member, Suresh Sundaram, Sr. Member, Arvind Easwaran, Sr. Member and Luigi Glielmo, Sr. Member Abstract—This paper presents a novel solution technique for Storage Units scheduling multi-energy system (MES) in a commercial urban PESS Power input to the electrical storage [kW]; building to perform price-based demand response and reduce SoCESS State of charge of the electrical storage [kW h]; energy costs. The MES scheduling problem is formulated as a δ mixed integer nonlinear program (MINLP), a non-convex NP- ESS charging/discharging status of the electrical stor- hard problem with uncertainties due to renewable generation age; and demand. A model predictive control approach is used PTES Cooling energy input to the thermal storage [kW]; to handle the uncertainties and price variations. This in-turn SoCTES State of charge of the thermal storage [kW h]; requires solving a time-coupled multi-time step MINLP dur- δTES charging/discharging status of the thermal storage; ing each time-epoch which is computationally intensive. This T ◦C investigation proposes an approach called the Scenario-Based TES Temperature inside the TES [ ]; Branch-and-Bound (SB3), a light-weight solver to reduce the Chiller Bank computational complexity. It combines the simplicity of convex −1 programs with the ability of meta-heuristic techniques to handle m_ C;j Cool water mass-flow rate of chiller j [kg h ]; complex nonlinear problems. The performance of the SB3 solver γC;j Binary ON-OFF status of chiller j; is validated in the Cleantech building, Singapore and the results PC Power consumed in the chiller bank [kW]; demonstrate that the proposed algorithm reduces energy cost by QC;j Thermal energy supplied by chiller j [kW]; about 17.26% and 22.46% as against solving a multi-time step heuristic optimization model.
    [Show full text]
  • The Big-M Method with the Numerical Infinite M
    Optimization Letters https://doi.org/10.1007/s11590-020-01644-6 ORIGINAL PAPER The Big-M method with the numerical infinite M Marco Cococcioni1 · Lorenzo Fiaschi1 Received: 30 March 2020 / Accepted: 8 September 2020 © The Author(s) 2020 Abstract Linear programming is a very well known and deeply applied field of optimization theory. One of its most famous and used algorithms is the so called Simplex algorithm, independently proposed by Kantoroviˇc and Dantzig, between the end of the 30s and the end of the 40s. Even if extremely powerful, the Simplex algorithm suffers of one initialization issue: its starting point must be a feasible basic solution of the problem to solve. To overcome it, two approaches may be used: the two-phases method and the Big-M method, both presenting positive and negative aspects. In this work we aim to propose a non-Archimedean and non-parametric variant of the Big-M method, able to overcome the drawbacks of its classical counterpart (mainly, the difficulty in setting the right value for the constant M). We realized such extension by means of the novel computational methodology proposed by Sergeyev, known as Grossone Methodol- ogy. We have validated the new algorithm by testing it on three linear programming problems. Keywords Big-M method · Grossone methodology · Infinity computer · Linear programming · Non-Archimedean numerical computing 1 Introduction Linear Programming (LP) is a branch of optimization theory which studies the min- imization (or maximization) of a linear objective function, subject to linear equality and/or linear inequality constraints. LP has found a lot of successful applications both in theoretical and real world contexts, especially in countless engineering applications.
    [Show full text]
  • Two Dimensional Search Algorithms for Linear Programming
    University of Nebraska at Omaha DigitalCommons@UNO Mathematics Faculty Publications Department of Mathematics 8-2019 Two dimensional search algorithms for linear programming Fabio Torres Vitor Follow this and additional works at: https://digitalcommons.unomaha.edu/mathfacpub Part of the Mathematics Commons Two dimensional search algorithms for linear programming by Fabio Torres Vitor B.S., Maua Institute of Technology, Brazil, 2013 M.S., Kansas State University, 2015 AN ABSTRACT OF A DISSERTATION submitted in partial fulfillment of the requirements for the degree DOCTOR OF PHILOSOPHY Department of Industrial and Manufacturing Systems Engineering Carl R. Ice College of Engineering KANSAS STATE UNIVERSITY Manhattan, Kansas 2019 Abstract Linear programming is one of the most important classes of optimization problems. These mathematical models have been used by academics and practitioners to solve numerous real world applications. Quickly solving linear programs impacts decision makers from both the public and private sectors. Substantial research has been performed to solve this class of problems faster, and the vast majority of the solution techniques can be categorized as one dimensional search algorithms. That is, these methods successively move from one solution to another solution by solving a one dimensional subspace linear program at each iteration. This dissertation proposes novel algorithms that move between solutions by repeatedly solving a two dimensional subspace linear program. Computational experiments demonstrate the potential of these newly developed algorithms and show an average improvement of nearly 25% in solution time when compared to the corresponding one dimensional search version. This dissertation's research creates the core concept of these two dimensional search algorithms, which is a fast technique to determine an optimal basis and an optimal solution to linear programs with only two variables.
    [Show full text]