Optimization by Simulated Annealing: an Experimental Evaluation; Part I

Total Page:16

File Type:pdf, Size:1020Kb

Optimization by Simulated Annealing: an Experimental Evaluation; Part I OPTIMIZATIONBY SIMULATEDANNEALING: AN EXPERIMENTAL EVALUATION;PART 1, GRAPH PARTITIONING DAVIDS. JOHNSON A T&T Bell Laboratories, Murray Hill, New Jersey CECILIAR. ARAGON University of California, Berkeley, California LYLE A. McGEOCH Amherst College, Amherst, Massachusetts CATHERINESCHEVON Johns Hopkins University, Baltimore, Maryland (ReceivedFebruary 1988; revision received January 1989; accepted February 1989) In this and two companionpapers, we reporton an extendedempirical study of the simulatedannealing approach to combinatorialoptimization proposed by S. Kirkpatricket al. That study investigatedhow best to adapt simulated annealingto particularproblems and comparedits performanceto that of more traditionalalgorithms. This paper(Part I) discusses annealing and our parameterizedgeneric implementationof it, describeshow we adapted this generic algorithmto the graphpartitioning problem, and reportshow well it comparedto standardalgorithms like the Kernighan- Lin algorithm.(For sparserandom graphs,it tended to outperformKernighan-Lin as the numberof verticesbecome large,even when its much greaterrunning time was taken into account.It did not performnearly so well, however,on graphsgenerated with a built-ingeometric structure.) We also discusshow we went aboutoptimizing our implementation, and describethe effectsof changingthe variousannealing parameters or varyingthe basicannealing algorithm itself. A ew approach to the approximate solution of new problems (even in the absence of deep insight difficult combinatorial optimization problems into the problems themselves) and, because of its recently has been proposed by Kirkpatrick,Gelatt and apparent ability to avoid poor local optima, it offers Vecchi (1983), and independently by Cerny (1985). hope of obtaining significantly better results. This simulatedannealing approach is based on ideas These observations, together with the intellectual from statistical mechanics and motivated by an anal- appeal of the underlying physical analogy, have ogy to the behavior of physical systems in the presence inspired articlesin the popular scientific press (Science of a heat bath. The nonphysicist, however, can view 82, 1982 and Physics Today 1982) as well as attempts it simply as an enhanced version of the familiar tech- to apply the approach to a variety of problems, in nique of local optimizationor iterativeimprovement, areas as diverse as VLSI design (Jepsen and in which an initial solution is repeatedly improved by Gelatt 1983, Kirkpatrick, Gelatt and Vecchi 1983, making small local alterationsuntil no such alteration Vecchi and Kirkpatrick 1983, Rowan and Hennessy yields a better solution. Simulated annealing random- 1985), pattern recognition (Geman and Geman 1984, izes this procedure in a way that allows for occasional Hinton, Sejnowski and Ackley 1984) and code gen- uphillmoves (changes that worsen the solution), in an eration (El Gamal et al. 1987), often with substantial attempt to reduce the probability of becoming stuck success. (See van Laarhoven and Aarts 1987 and in a poor but locally optimal solution. As with local Collins, Eglese and Golden 1988 for more up-to-date search, simulated annealing can be adapted readily to and extensive bibliographiesof applications.)Many of Subject classifications: Networks/graphs, heuristics: algorithms for graph partitioning. Simulation, applications: optimization by simulated annealing. Operations Research 0030-364X/89/3706-0865 $01.25 Vol. 37, No. 6, November-December 1989 865 ? 1989 Operations Research Society of America This content downloaded from 128.208.219.145 on Thu, 04 Jun 2015 21:20:57 UTC All use subject to JSTOR Terms and Conditions 866 / JOHNSON ET AL. the practical applications of annealing, however, have for sparse random graphs it tended to outperform been in complicated problem domains, where pre- Kernighan-Lin as the number of vertices became vious algorithms either did not exist or perfom-led large. For a class of random graphs with built-in quite poorly. In this paper and its two companions, geometric structure,however, Kernighan-Linwon the we investigate the performanceof simulated annealing comparisons by a substantialmargin. Thus, simulated in more competitive arenas, in the hope of obtaining annealing's success can best be described as mixed. a better view of the ultimate value and limitations of Section 4 describes the experiments by which we the approach. optimized the annealing parametersused to generate The arena for this paper is the problem of partition- the results reportedin Section 3. Section 5 investigates ing the vertices of a graph into two equal size sets to the effectiveness of various modifications and alter- minimize the number of edges with endpoints in natives to the basic annealing algorithm. Section 6 both sets. This application was first proposed by discusses some of the other algorithms that have been Kirkpatrick, Gelatt and Vecchi, but was not exten- proposed for graph partitioning, and considers how sively studied there. (Subsequently, Kirkpatrick 1984 these might factor into our comparisons. We conclude went into the problem in more detail, but still did not in Section 7 with a summary of our observationsabout deal adequately with the competition.) the value of simulated annealing for the graph parti- Our paper is organized as follows. In Section 1, we tioning problem, and with a list of lessons learned that introduce the graph partitioning problem and use it may well be applicable to implementations of simu- to illustrate the simulated annealing approach. We lated annealing for other combinatorial optimization also sketch the physical analogy on which annealing problems. is based, and discuss some of the reasons for optimism In the two companion papers to follow, we will (and for skepticism) concerning it. Section 2 presents report on our attempts to apply these lessons to three the details of our implementation of simulated anneal- other well studied problems: Graph Coloring and ing, describing a parameterized, generic annealing Number Partitioning (Johnson et al 1990a), and the algorithm that calls problem-specificsubroutines, and Traveling Salesman Problem (Johnson et al. 1990b). hence, can be used in a variety of problem domains. Sections 3 through 6 present the results of our 1. SIMULATEDANNEALING: THE BASIC experiments with simulated annealing on the graph CONCEPTS partitioning problem. Comparisons between anneal- ing and its rivals are made difficult by the fact that 1.1. Local Optimization the performance of annealing depends on the partic- To understand simulated annealing, one must first ular annealing schedule chosen and on other, more understand local optimization. A combinatorial opti- problem-specific parameters. Methodological ques- mization problem can be specified by identifying a set tions also arise because annealing and its main com- of solutions together with a cost function that assigns petitors are randomized algorithms (and, hence, can a numerical value to each solution. An optimal solu- give a variety of answers for the same instance) and tion is a solution with the minimum possible cost because they have running times that differ by factors (there may be more than one such solution). Given as large as 1,000 on our test instances. Thus, if com- an arbitrarysolution to such a problem, local opti- parisons are to be convincing and fair, they must be mization attempts to improve on that solution by a based on large numbers of independent runs of the series of incremental, local changes. To define a local algorithms, and we cannot simply compare the aver- optimization algorithm, one first specifies a method age cutsizes found. (In the time it takes to perform for perturbingsolutions to obtain different ones. The one run of the slower algorithm, one could perform set of solutions that can be obtained in one such step many runs of the faster one and take the best solution from a given solution A is called the neighborhoodof found.) A. The algorithmthen performsthe simple loop shown Section 3 describes the problem-specific details of in Figure 1, with the specific methods for choosing S our implementation of annealing for graph partition- and S' left as implementation details. ing. It then introduces two general types of test graphs, Although S need not be an optimal solution when and summarizes the results of our comparisons the loop is finally exited, it will be locally optimal in between annealing, local optimization, and an algo- that none of its neighbors has a lower cost. The hope rithm due to Kernighan-Lin(1970) that has been the is that locally optimal will be good enough. long-reigning champion for this problem. Annealing To illustratethese concepts, let us consider the graph almost always outperformed local optimization, and partitioning problem that is to be the topic of Section This content downloaded from 128.208.219.145 on Thu, 04 Jun 2015 21:20:57 UTC All use subject to JSTOR Terms and Conditions GraphPartitioning by SimulatedAnnealing / 867 1. Get an initialsolution S. Simulated annealing is an approach that attempts 2. While there is an untested neighbor of S do the to avoid entrapment in poor local optima by allowing following. an occasional uphill move. This is done under the 2.1 Let S' be an untested neighbor of S. influence of a random number generatorand a control 2.2 If cost (S') < cost (S), set S = S'. parametercalled the temperature.As typically imple- 3. Return S. mented, the simulated annealing approach involves a pair of nested loops and
Recommended publications
  • Simulated Annealing Hyper-Heuristic for a Shelf Space Allocation on Symmetrical Planograms Problem
    S S symmetry Article Simulated Annealing Hyper-Heuristic for a Shelf Space Allocation on Symmetrical Planograms Problem Kateryna Czerniachowska * and Marcin Hernes Department of Process Management, Wroclaw University of Economics and Business, Komandorska 118/120, 53-345 Wroclaw, Poland; [email protected] * Correspondence: [email protected] Abstract: The allocation of products on shelves is an important issue from the point of view of effective decision making by retailers. In this paper, we investigate a practical shelf space allocation model which takes into account the number of facings, capping, and nesting of a product. We divide the shelf into the segments of variable size in which the products of the specific types could be placed. The interconnections between products are modelled with the help of categorizing the products into specific types as well as grouping some of them into clusters. This results in four groups of constraints—shelf constraints, shelf type constraints, product constraints, position allocation constraints—that are used in the model for aesthetic symmetry of a planogram. We propose a simulated annealing algorithm with improvement and reallocation procedures to solve the planogram profit maximization problem. Experiments are based on artificial data sets that have been generated according to real-world conditions. The efficiency of the designed algorithm has been estimated using the CPLEX solver. The computational tests demonstrate that the proposed algorithm gives valuable results in an acceptable time. Citation: Czerniachowska, K.; Hernes, M. Simulated Annealing Keywords: retailing decisions; shelf space allocation; simulated annealing; heuristics; merchandising Hyper-Heuristic for a Shelf Space Allocation on Symmetrical Planograms Problem.
    [Show full text]
  • A Simulated Annealing Based Inexact Oracle for Wasserstein Loss Minimization
    A Simulated Annealing Based Inexact Oracle for Wasserstein Loss Minimization Jianbo Ye 1 James Z. Wang 1 Jia Li 2 Abstract ryl(x; ·) is computed in O(m) time, m being the complex- ity of outcome variables x or y. This part of calculation is Learning under a Wasserstein loss, a.k.a. Wasser- often negligible compared with the calculation of full gra- stein loss minimization (WLM), is an emerging dient with respect to the model parameters. But this is no research topic for gaining insights from a large longer the case in learning problems based on Wasserstein set of structured objects. Despite being concep- distance due to the intrinsic complexity of the distance. tually simple, WLM problems are computation- We will call such problems Wasserstein loss minimization ally challenging because they involve minimiz- (WLM). Examples of WLMs include Wasserstein barycen- ing over functions of quantities (i.e. Wasserstein ters (Li & Wang, 2008; Agueh & Carlier, 2011; Cuturi & distances) that themselves require numerical al- Doucet, 2014; Benamou et al., 2015; Ye & Li, 2014; Ye gorithms to compute. In this paper, we intro- et al., 2017b), principal geodesics (Seguy & Cuturi, 2015), duce a stochastic approach based on simulated nonnegative matrix factorization (Rolet et al., 2016; San- annealing for solving WLMs. Particularly, we dler & Lindenbaum, 2009), barycentric coordinate (Bon- have developed a Gibbs sampler to approximate neel et al., 2016), and multi-label classification (Frogner effectively and efficiently the partial gradients of et al., 2015). a sequence of Wasserstein losses. Our new ap- proach has the advantages of numerical stability Wasserstein distance is defined as the cost of matching two and readiness for warm starts.
    [Show full text]
  • University of Waterloo Faculty of Engineering
    University of Waterloo Faculty of Engineering Global Optimization of General Failure Surfaces in Slope Analysis by Hybrid Simulated Annealing Rocscience, Inc. Toronto, Ontario Prepared by Su, Xiao ID 20251755 2B Chemical Engineering May 1st , 2009 4105 Credit Pointe Dr, Mississauga, Ontario L5M 3K7 May 1st, 2009 Dr Thomas Duever, Chair of Chemical Engineering University of Waterloo Waterloo, Ontario N2L 3G1 Dear Dr Duever, I am submitting this document, entitled “Global Optimization of General Failure Surfaces in Slope Analysis by Hybrid Simulated Annealing” as my 2B Work Report for Rocscience, Inc. My project investigated the efficiency of a hybrid simulated annealing (HSA) method for finding the minima (critical failure surfaces and factors of safety) of slope stability equations. The HSA combines a Very Fast Simulated Annealing (VFSA) and a Local Monte-Carlo (LMC) search, the latter being an optimization algorithm designed specifically for solving slope stability problems. This work is an extension as well as a generalization of the project from my summer work-term (2008), also undertaken at Rocscience. During both terms, I worked under the supervision of Dr. Brent Corkum, software development manager at Rocscience. I implemented the simulated annealing algorithm as a new optimization method in SLIDE 5.0, a slope stability analysis program developed by Rocscience. In addition to learning professional C++ and principles of object oriented programming, I had the opportunity to study the theoretical and practical aspects of non-convex optimization, especially the performance of stochastic methods when used on multi- modal functions. I also learned soil mechanics and slope stability principles. This report was written entirely by me and has not received any previous academic credit at Rocscience or the University of Waterloo.
    [Show full text]
  • Arxiv:2101.10154V1 [Cond-Mat.Dis-Nn] 25 Jan 2021
    Variational Neural Annealing 1, 2, 3, 1 1, 2 2, 3 1, 2 Mohamed Hibat-Allah, ∗ Estelle M. Inack, Roeland Wiersema, Roger G. Melko, and Juan Carrasquilla 1Vector Institute, MaRS Centre, Toronto, Ontario, M5G 1M1, Canada 2Department of Physics and Astronomy, University of Waterloo, Ontario, N2L 3G1, Canada 3Perimeter Institute for Theoretical Physics, Waterloo, ON N2L 2Y5, Canada (Dated: January 26, 2021) Many important challenges in science and technology can be cast as optimization problems. When viewed in a statistical physics framework, these can be tackled by simulated annealing, where a gradual cooling procedure helps search for groundstate solutions of a target Hamiltonian. While powerful, simulated annealing is known to have prohibitively slow sampling dynamics when the optimization landscape is rough or glassy. Here we show that by generalizing the target distribution with a parameterized model, an analogous annealing framework based on the variational principle can be used to search for groundstate solutions. Modern autoregressive models such as recurrent neural networks provide ideal parameterizations since they can be exactly sampled without slow dynamics even when the model encodes a rough landscape. We implement this procedure in the classical and quantum settings on several prototypical spin glass Hamiltonians, and find that it significantly outperforms traditional simulated annealing in the asymptotic limit, illustrating the potential power of this yet unexplored route to optimization. P I. INTRODUCTION <latexit sha1_base64="Wk+wWdaEIxpmR5GD8uDFICvy2Ig=">AAACNHicfVDLSgMxFL1TX7W+Rl26CRbBVZkRQZdFN4KbCvYBbRkyaaYNzWSGJGMpQz/KjR/iRgQXirj1G8y0s6iteCBwOOfcJPf4MWdKO86rVVhZXVvfKG6WtrZ3dvfs/YOGihJJaJ1EPJItHyvKmaB1zTSnrVhSHPqcNv3hdeY3H6hULBL3ehzTboj7ggWMYG0kz76teSnq9KKRwFJGI9RJ4kUyZ/6Xm5Q8u+xUnCnQMnFzUoYcNc9+NjeSJKRCE46VartOrLsplpoRTielTqJojMkQ92nbUIFDqrrpdOkJOjFKDwWRNEdoNFXnJ1IcKjUOfZMMsR6oRS8T//LaiQ4uuykTcaKpILOHgoQjHaGsQdRjkhLNx4ZgIpn5KyIDLDHRpuesBHdx5WXSOKu4TsW9Oy9Xr/I6inAEx3AKLlxAFW6gBnUg8Agv8A4f1pP1Zn1aX7NowcpnDuEXrO8fzwWstw==</latexit>
    [Show full text]
  • A Simulated Annealing Algorithm for Solving Two-Echelon Vehicle Routing Problem with Locker Facilities
    algorithms Article A Simulated Annealing Algorithm for Solving Two-Echelon Vehicle Routing Problem with Locker Facilities A. A. N. Perwira Redi 1 , Parida Jewpanya 2,*, Adji Candra Kurniawan 3, Satria Fadil Persada 4, Reny Nadlifatin 5 and Oki Anita Candra Dewi 6 1 Industrial Engineering Department, BINUS Graduate Program—Master of Industrial Engineering, Bina Nusantara University, Jakarta 11480, Indonesia; [email protected] 2 Department of Industrial Engineering, Rajamangala University of Technology Lanna Tak, Tak 63000, Thailand 3 Department of Logistics Engineering, Universitas Pertamina, Jakarta 12220, Indonesia; [email protected] 4 Department of Business Management, Institut Teknologi Sepuluh Nopember, Surabaya 60111, Indonesia; [email protected] 5 Department of Technology Management, Institut Teknologi Sepuluh Nopember, Surabaya 60111, Indonesia; [email protected] 6 Logistics Engineering Department, Universitas Internasional Semen Indonesia, Gresik 61122, Indonesia; [email protected] * Correspondence: [email protected] Received: 6 August 2020; Accepted: 1 September 2020; Published: 3 September 2020 Abstract: We consider the problem of utilizing the parcel locker network for the logistics solution in the metropolitan area. Two-echelon distribution systems are attractive from an economic standpoint, whereas the product from the depot can be distributed from or to intermediate facilities. In this case, the intermediate facilities are considered as locker facilities present in an accessible location in the vicinity of the final customers. In addition, the utilization of locker facilities can reduce the cost caused by the unattended deliveries. The problem is addressed as an optimization model that formulated into an integer linear programming model denoted as the two-echelon vehicle routing problem with locker facilities (2EVRP-LF).
    [Show full text]
  • Microwave Imaging for Conducting Scatterers by Hybrid Particle Swarm Optimization with Simulated Annealing
    2011 8th International Multi-Conference on Systems, Signals & Devices MICROWAVE IMAGING FOR CONDUCTING SCATTERERS BY HYBRID PARTICLE SWARM OPTIMIZATION WITH SIMULATED ANNEALING Bouzid Mhamdi, Khaled Grayaa and Taoufik Aguili Syscom Laboratory, Engineer School of Tunis - BP 37 Belvedere, Tunis - 1002, Tunisia [email protected] ABSTRACT The first is based on gradient searching schemes such as the Newton-Kantorovitch method [3], modified In this paper, a microwave imaging technique for gradient method [4] and Levenberg-Marguart algorithm reconstructing the shape of two-dimensional perfectly [5]. It is well-known that these deterministic techniques conducting scatterers by means of a stochastic used for fast reconstruction of microwave images tend to optimization approach is investigated. get trapped in a local extreme and suffer from a major Based on the boundary condition and the measured drawback, where the final image is highly dependent on scattered field derived by transverse magnetic the initial trial solution. To overcome this obstacle, a illuminations, a set of nonlinear integral equations is second class of population based stochastic methods such obtained and the imaging problem is reformulated in to as the genetic algorithm (GA) [1-2], simulated annealing an optimization problem. (SA) [6] and neural network (NN) [7] has become A hybrid approximation algorithm, called PSO-SA, is attractive alternatives to reconstruct microwave images. developed in this work to solve the scattering inverse These techniques consider the imaging problem as a problem. In the hybrid algorithm, particle swarm optimization (PSO) combines global search and local global optimization problem, and reconstruct the correct search for finding the optimal results assignment with image by searching for the optimal solution through the reasonable time and simulated annealing (SA) uses use of either rivalry or cooperation strategies in the certain probability to avoid being trapped in a local problem space.
    [Show full text]
  • Arxiv:1904.04123V2 [Stat.ML] 10 Oct 2019 Human Efforts Invested in finding Good Network Architec- Pected final Architecture, and Choose the Components That Tures
    ASAP: Architecture Search, Anneal and Prune Asaf Noy1, Niv Nayman1, Tal Ridnik1, Nadav Zamir1, Sivan Doveh2, Itamar Friedman1, Raja Giryes2, and Lihi Zelnik-Manor1 1DAMO Academy, Alibaba Group (fi[email protected]) 2School of Electrical Engineering, Tel-Aviv University Tel-Aviv, Israel Abstract methods, such as ENAS [34] and DARTS [12], reduced the search time to a few GPU days, while not compromising Automatic methods for Neural Architecture Search much on accuracy. While this is a major advancement, there (NAS) have been shown to produce state-of-the-art net- is still a need to speed up the process to make the automatic work models. Yet, their main drawback is the computa- search affordable and applicable to more problems. tional complexity of the search process. As some primal In this paper we propose an approach that further re- methods optimized over a discrete search space, thousands duces the search time to a few hours, rather than days or of days of GPU were required for convergence. A recent weeks. The key idea that enables this speed-up is relaxing approach is based on constructing a differentiable search the discrete search space to be continuous, differentiable, space that enables gradient-based optimization, which re- and annealable. For continuity and differentiability, we fol- duces the search time to a few days. While successful, it low the approach in DARTS [12], which shows that a con- still includes some noncontinuous steps, e.g., the pruning of tinuous and differentiable search space allows for gradient- many weak connections at once. In this paper, we propose a based optimization, resulting in orders of magnitude fewer differentiable search space that allows the annealing of ar- computations in comparison with black-box search, e.g., chitecture weights, while gradually pruning inferior opera- of [48, 35].
    [Show full text]
  • Lecture 19 -- Simulated Annealing
    Simulated Annealing Biostatistics 615/815 Lecture 19 Scheduling z Need to pick a date for mid-term z Default date is December 20, 2006 z We could have it earlier… • For example, on December 12, 2006? z What do you prefer? So far … z “Greedy” optimization methods • Can get trapped at local minima • Outcome might depend on starting point z Examples: • Golden Search • Nelder-Mead Simplex Optimization • E-M Algorithm Today … z Simulated Annealing z Markov-Chain Monte-Carlo method z Designed to search for global minimum among many local minima The Problem z Most minimization strategies find the nearest local minimum z Standard strategy • Generate trial point based on current estimates • Evaluate function at proposed location • Accept new value if it improves solution The Solution z We need a strategy to find other minima z This means, we must sometimes select new points that do not improve solution z How? Annealing z One manner in which crystals are formed z Gradual cooling of liquid … • At high temperatures, molecules move freely • At low temperatures, molecules are "stuck" z If cooling is slow • Low energy, organized crystal lattice formed Simulated Annealing z Analogy with thermodynamics z Incorporate a temperature parameter into the minimization procedure z At high temperatures, explore parameter space z At lower temperatures, restrict exploration Markov Chain z Start with some sample • A set of mixture parameters z Propose a change • Edit mixture parameters in some way z Decide whether to accept change • Decision is based on relative probabilities
    [Show full text]
  • Simulated Annealing
    Appendix A Simulated Annealing A.1 Introduction Simulated annealing can be considered as an extension of local optimization meth- ods because: • This approach can be applied to criteria that are neither continuous nor con- tinuously differentiable. It is just necessary to be able to compute the value of the criterion for any feasible solution. Thus, the criterion may be given by any type of function or even by an algorithm that returns a numerical value starting from the values of the parameters that define a solution. • The variables can be of a qualitative nature, needing only to be able to derive from their “values” a quantitative value of the criterion. Indeed, simulated annealing easily applies to combinatorial optimization prob- lems. A problem P that belongs to the field of application of simulated annealing is expressed as any optimization problem: Find *∈ Ss such that )*( = Opt sfsf )( ∈Ss where: • S is the set of feasible solutions. A feasible solution is one that satisfies all the constraints. We will see some examples in the next section. • f is a “function” in the broadest sense, as explained above. This function is the criterion. • Opt refers either to minimization or to maximization, depending on the type of problem. 450 A Simulated Annealing A.2 Basic Requirements In order to apply simulated annealing, we must be able to: • Compute the value of the criterion for any feasible solution. • Define an initial feasible solution. • Derive a neighboring feasible solution from any current solution. The criterion depends on the problem to be solved. An initial feasible solution may be difficult to find.
    [Show full text]
  • Heuristics: Simulated Annealing and Genetic Algorithms
    Simulated annealing and genetic algorithms CE 377K Stephen D. Boyles Spring 2015 1 Introduction Different optimization problems must be solved in different ways. Even just for the simple examples developed at the start of this course, the way that one solves the shortest path problem, for instance, is very different from the way that one solves the transit frequency setting problem. Later in this course, we will learn examples of these types of specific solution methods. But first, we start with a broad class of solution methods which can be applied to nearly any optimization problem. In optimization, there is often a tradeoff between how widely applicable a solution method is, and how efficient or effective it is at solving specific problems. For any one specific problem, a tailor- made solution process is likely much faster than a generally-applicable method, but generating such a method requires more effort and specialized knowledge, and is less \rewarding" in the sense that the method can only be more narrowly applied. So, the first solution methods you will see in this course are very widely applicable. They are examples of heuristics, which are not guaranteed to find the global optimum solution, but tend to work reasonably well in practice. In practice, they only tend to be applied for very large or very complicated problems which cannot be solved exactly in a reasonable amount of time with our current knowledge. But until you learn the more specialized solution methods later in the class, you can also use these even for smaller or simpler optimization problems too.
    [Show full text]
  • Improving Simulated Annealing Through Derandomization
    Improving Simulated Annealing through Derandomization Mathieu Gerber∗ Luke Bornn Department of Statistics Harvard University We propose and study a version of simulated annealing (SA) on continu- ¯ ous state spaces based on (t; s)R-sequences. The parameter R 2 N regulates the degree of randomness of the input sequence, with the case R = 0 corres- ponding to IID uniform random numbers and the limiting case R = 1 to (t; s)-sequences. Our main result, obtained for rectangular domains, shows that the resulting optimization method, which we refer to as QMC-SA, con- verges almost surely to the global optimum of the objective function ' for any R 2 N. When ' is univariate, we are in addition able to show that the completely deterministic version of QMC-SA is convergent. A key property of these results is that they do not require objective-dependent conditions on the cooling schedule. As a corollary of our theoretical analysis, we provide a new almost sure convergence result for SA which shares this property under minimal assumptions on '. We further explain how our results in fact apply to a broader class of optimization methods including for example threshold accepting, for which to our knowledge no convergence results currently ex- ist. We finally illustrate the superiority of QMC-SA over SA algorithms in a numerical study. Keywords: Global optimization; Quasi-Monte Carlo; Randomized quasi- Monte Carlo; Simulated annealing; Threshold accepting 1. Introduction arXiv:1505.03173v4 [stat.CO] 5 Sep 2016 Simulated annealing (SA) belongs to the class of stochastic optimization techniques, a popular suite of tools to find the global optimum of multi-modal and/or non-differentiable functions defined on continuous sets (Geman and Hwang, 1986).
    [Show full text]
  • Improved Salp Swarm Algorithm with Simulated Annealing for Solving Engineering Optimization Problems
    S S symmetry Article Improved Salp Swarm Algorithm with Simulated Annealing for Solving Engineering Optimization Problems Qing Duan, Lu Wang, Hongwei Kang * , Yong Shen, Xingping Sun and Qingyi Chen School of Software, Yunnan University, Kunming 650000, China; [email protected] (Q.D.); [email protected] (L.W.); [email protected] (Y.S.); [email protected] (X.S.); [email protected] (Q.C.) * Correspondence: [email protected] Abstract: Swarm-based algorithm can successfully avoid the local optimal constraints, thus achieving a smooth balance between exploration and exploitation. Salp swarm algorithm (SSA), as a swarm- based algorithm on account of the predation behavior of the salp, can solve complex daily life optimization problems in nature. SSA also has the problems of local stagnation and slow convergence rate. This paper introduces an improved salp swarm algorithm, which improve the SSA by using the chaotic sequence initialization strategy and symmetric adaptive population division. Moreover, a simulated annealing mechanism based on symmetric perturbation is introduced to enhance the local jumping ability of the algorithm. The improved algorithm is referred to SASSA. The CEC standard benchmark functions are used to evaluate the efficiency of the SASSA and the results demonstrate that the SASSA has better global search capability. SASSA is also applied to solve engineering optimization problems. The experimental results demonstrate that the exploratory and exploitative proclivities of the proposed algorithm and its convergence patterns are vividly improved. Citation: Duan, Q.; Wang, L.; Kang, H.; Shen, Y.; Sun, X.; Chen, Q. Keywords: swarm-based algorithm; salp swarm algorithm; single objective optimization; symmetric Improved Salp Swarm Algorithm perturbation; simulated annealing; engineering optimization problems with Simulated Annealing for Solving Engineering Optimization Problems.
    [Show full text]