On the Design of (Meta)Heuristics

Total Page:16

File Type:pdf, Size:1020Kb

On the Design of (Meta)Heuristics On the design of (meta)heuristics Tony Wauters CODeS Research group, KU Leuven, Belgium There are two types of people • Those who use heuristics • And those who don’t ☺ 2 Tony Wauters, Department of Computer Science, CODeS research group Heuristics Origin: Ancient Greek: εὑρίσκω, "find" or "discover" Oxford Dictionary: Proceeding to a solution by trial and error or by rules that are only loosely defined. Tony Wauters, Department of Computer Science, CODeS research group 3 Properties of heuristics + Fast* + Scalable* + High quality solutions* + Solve any type of problem (complex constraints, non-linear, …) - Problem specific (but usually transferable to other problems) - Cannot guarantee optimallity * if well designed 4 Tony Wauters, Department of Computer Science, CODeS research group Heuristic types • Constructive heuristics • Metaheuristics • Hybrid heuristics • Matheuristics: hybrid of Metaheuristics and Mathematical Programming • Other hybrids: Machine Learning, Constraint Programming,… • Hyperheuristics 5 Tony Wauters, Department of Computer Science, CODeS research group Constructive heuristics • Incrementally construct a solution from scratch. • Easy to understand and implement • Usually very fast • Reasonably good solutions 6 Tony Wauters, Department of Computer Science, CODeS research group Metaheuristics A Metaheuristic is a high-level problem independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms. Instead of reinventing the wheel when developing a heuristic for each different problem, metaheuristics offer several tools that can be used as the skeleton for the development of a heuristic. 7 Tony Wauters, Department of Computer Science, CODeS research group Metaheuristics • Key property: iteratively improve one or more solutions • Can be combined with constructive heuristics (initial solution) • Local search based • Steepest descent / Hill climbing • Tabu search, Variable Neighbourhood Search (VNS) • Simulated Annealing, Late Acceptance • GRASP, Guided Local Search • Population based • Genetic algorithms / memetic algorithms • Ant Colony Optimization • PSO (Particle swarm optimization) 8 Tony Wauters, Department of Computer Science, CODeS research group Hill climbing • Variants • Stochastic Hill-climbing • First-improving 9 Tony Wauters, Department of Computer Science, CODeS research group Late acceptance Hill-climbing [Burke and Bykov 2017] Tony Wauters, Department of Computer Science, CODeS research group 10 Metaheuristics the metaphor exposed Slides based on: Sörensen, K. Metaheuristics – the metaphor exposed. International Transactions in Operational Research, 22: 3-18, 2015. Thanks to Everton for sharing these slides! 12 Tony Wauters, Department of Computer Science, CODeS research group History • Early research on heuristics focused on human intuition, insight and learning as a way of solving problems. • Metaheuristics followed the same strategy. • Example: Tabu search • Memorizes part of the trajectory the search had taken in the recent and not-so-recent past. • Similar to a person looking for water in the desert trying to avoid running in circles. 13 Tony Wauters, Department of Computer Science, CODeS research group History • Eventually new methods were based on processes that had little to do with human intuition. • Simulated annealing • Natural evolution • Ant-colony optimization 14 Tony Wauters, Department of Computer Science, CODeS research group History • Simulated annealing • One of the earliest metaphor-based heuristics. • Analogy to the thermodynamic process of annealing. • Attempt to settle atoms into the lowest possible energy state. • Total energy state of the atoms quality of the solution. • Change of atom positions changes in the solution (moves). 15 Tony Wauters, Department of Computer Science, CODeS research group History • Natural evolution • Probably the most successful metaphor used until now. • Individuals in a population competing for survival. • Individuals or chromosomes solutions. • Fitness objective function value. • Crossover / Mutation changes in the solution. 16 Tony Wauters, Department of Computer Science, CODeS research group History • Ant-colony optimization • Models the optimization process upon the behavior of a colony of ants searching for food. • Agents (ants) construct solutions in parallel. • Each element of each solution has its own measure • pheromone • Elements with stronger pheromones have higher probability of being selected for future solutions. Tony Wauters, Department of Computer Science, CODeS research group 17 History • With the increase in number of metaheuristics being created comes the question: Until what point creating metaphors would result in really effective approaches? 18 Tony Wauters, Department of Computer Science, CODeS research group Not so good metaphors 19 Tony Wauters, Department of Computer Science, CODeS research group Metaphors “Models of nature that are relied upon for inspiration [in creating new metaheuristics] are ubiquitous, and it is easy to conjure up examples whose metaphorical possibilities have not yet been tapped.[…] It is perhaps surprising in retrospect that the behavior of bees has not been selected as a basis for one of the ‘new’ problem solving methods.” – Glover & Laguna (1998) 20 Tony Wauters, Department of Computer Science, CODeS research group Metaphors • People took seriously what he meant as sarcasm: • Bees optimization algorithm (2004). • Bee colony algorithm (2005). • Bee mating optimization algorithm (2006). • Difference between all insect algorithms proved marginal at best. 21 Tony Wauters, Department of Computer Science, CODeS research group “Novel” metaphor-based method • Harmony search metaheuristic (2001) • Changes the terminology of optimization into a musical one. • Harmony solution. • Note decision variable. • Sounds better has a better objective function. • Harmony memory set of solutions. • Concept: A musician playing or composing music has to “draw notes from a uniform random distribution”. 22 Tony Wauters, Department of Computer Science, CODeS research group Harmony search • General idea: • generates a set of random initial solutions; • tries to find better solutions by combining existing ones; • changes the values of decision variables in some of the solutions in the set. • In the end it is just another evolutionary algorithm… 23 Tony Wauters, Department of Computer Science, CODeS research group Harmony search • “Most importantly, when I searched Wikipedia, I could not find the structure which equals Harmony Search”. • HS was originally “created” to solve the design of water distribution networks. • 3 or 4 instances are tested. • A simple constructive heuristic outperformed the method. • Google scholar search for “harmony search”: • 2010 – around 500 results. • 2012 – around 3000 results • 2 hours ago – around 23000 results 24 Tony Wauters, Department of Computer Science, CODeS research group How are these heuristics accepted? • Fetish with novelty. • Race to make new scientific discoveries. • Most common is for the method to be marketed by a single researcher: • Co-authors a sizeable fraction of all papers on this method. • Social media website profiles for these methods: • Intelligent water drops algorithm officially “likes” the galaxy-based search algorithm. 25 Tony Wauters, Department of Computer Science, CODeS research group How are these heuristics accepted? • Up-the-wall game (Burke et al., 2009) • Game with no rules, just a goal: • To obtain better results than your opponents • Papers on metaheuristics are not accepted if they contain only an interesting insight • The methods presented need to be successful players in the up-the-wall game. • Methods have to be “tuned” for an instance-specific basis. 26 Tony Wauters, Department of Computer Science, CODeS research group Metaheuristic deconstruction • Which components it consists of? • Which other metaheuristics these components appear? • How to adapt this components to a specific problem that is being solved? • Please, use the general terminology • Call a solution as solution. 27 Tony Wauters, Department of Computer Science, CODeS research group Design of (meta)heuristics 28 Tony Wauters, Department of Computer Science, CODeS research group Design of heuristic techniques I • Solution representation • Direct vs indirect representation • Often a permutation • Datastructures • N-dimensional array • Tree, graph • Linked List, Queue, Stack 29 Tony Wauters, Department of Computer Science, CODeS research group Design of heuristic techniques II • How can we modify our solutions (and make them better)? • Solution modifications (aka the moves) • For example: insert, remove, swap, shift, rotate, .. • Depends on solution representation. • Important points to consider: • Maximum coverage of the solution space • Connectivity • Some basic Landscape analysis might help 30 Tony Wauters, Department of Computer Science, CODeS research group Intensification vs diversification • Intensification vs diversification [Glover and Laguna 1997] • In intensification, the promising regions are explored more thoroughly in the hope to find better solutions. • In diversification, non-explored must be visited to be sure that all regions of the search space are evenly explored and that the search is not confined to only a reduced number of regions. 31 Tony Wauters, Department of Computer Science, CODeS research group Key Ingredients • Greediness • Randomness • Memory (e.g. Tabu Search) • Adaptiveness • Other things
Recommended publications
  • Hybridizations of Metaheuristics with Branch & Bound Derivates
    Hybridizations of Metaheuristics With Branch & Bound Derivates Christian Blum1, Carlos Cotta2, Antonio J. Fern´andez2,Jos´e E. Gallardo2, and Monaldo Mastrolilli3 1 ALBCOM research group Universitat Polit`ecnica de Catalunya [email protected] 2 Dept. Lenguajes y Ciencias de la Computaci´on Universidad de M´alaga {ccottap,afdez,pepeg}@lcc.uma.es 3 Istituto Dalle Molle di Studi sull’Intelligenza Artificiale (IDSIA) [email protected] Summary. An important branch of hybrid metaheuristics concerns the hybridiza- tion with branch & bound derivatives. In this chapter we present examples for two different types of hybridization. The first one concerns the use of branch & bound fea- tures within construction-based metaheuristics in order to increase their efficiancy. The second example deals with the use of a metaheuristic, in our case a memetic algorithm, in order to increase the efficiancy of branch & bound, respectively branch & bound derivatives such as beam search. The quality of the resulting hybrid tech- niques is demonstrated by means of the application to classical string problems: the longest common subsequence problem and the shortest common supersequence problem. 1 Introduction One of the basic ingredients of an optimization technique is a mechanism for exploring the search space, that is, the space of valid solutions to the con- sidered optimization problem. Algorithms belonging to the important class of constructive optimization techniques tackle an optimization problem by ex- ploring the search space in form of a tree, a so-called search tree.Thesearch tree is generally defined by an underlying solution construction mechanism. Each path from the root node of the search tree to one of the leaves corre- sponds to the process of constructing a candidate solution.
    [Show full text]
  • Metaheuristics1
    METAHEURISTICS1 Kenneth Sörensen University of Antwerp, Belgium Fred Glover University of Colorado and OptTek Systems, Inc., USA 1 Definition A metaheuristic is a high-level problem-independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms (Sörensen and Glover, To appear). Notable examples of metaheuristics include genetic/evolutionary algorithms, tabu search, simulated annealing, and ant colony optimization, although many more exist. A problem-specific implementation of a heuristic optimization algorithm according to the guidelines expressed in a metaheuristic framework is also referred to as a metaheuristic. The term was coined by Glover (1986) and combines the Greek prefix meta- (metá, beyond in the sense of high-level) with heuristic (from the Greek heuriskein or euriskein, to search). Metaheuristic algorithms, i.e., optimization methods designed according to the strategies laid out in a metaheuristic framework, are — as the name suggests — always heuristic in nature. This fact distinguishes them from exact methods, that do come with a proof that the optimal solution will be found in a finite (although often prohibitively large) amount of time. Metaheuristics are therefore developed specifically to find a solution that is “good enough” in a computing time that is “small enough”. As a result, they are not subject to combinatorial explosion – the phenomenon where the computing time required to find the optimal solution of NP- hard problems increases as an exponential function of the problem size. Metaheuristics have been demonstrated by the scientific community to be a viable, and often superior, alternative to more traditional (exact) methods of mixed- integer optimization such as branch and bound and dynamic programming.
    [Show full text]
  • AI-KI-B Heuristic Search
    AI-KI-B Heuristic Search Ute Schmid & Diedrich Wolter Practice: Johannes Rabold Cognitive Systems and Smart Environments Applied Computer Science, University of Bamberg last change: 3. Juni 2019, 10:05 Outline Introduction of heuristic search algorithms, based on foundations of problem spaces and search • Uniformed Systematic Search • Depth-First Search (DFS) • Breadth-First Search (BFS) • Complexity of Blocks-World • Cost-based Optimal Search • Uniform Cost Search • Cost Estimation (Heuristic Function) • Heuristic Search Algorithms • Hill Climbing (Depth-First, Greedy) • Branch and Bound Algorithms (BFS-based) • Best First Search • A* • Designing Heuristic Functions • Problem Types Cost and Cost Estimation • \Real" cost is known for each operator. • Accumulated cost g(n) for a leaf node n on a partially expanded path can be calculated. • For problems where each operator has the same cost or where no information about costs is available, all operator applications have equal cost values. For cost values of 1, accumulated costs g(n) are equal to path-length d. • Sometimes available: Heuristics for estimating the remaining costs to reach the final state. • h^(n): estimated costs to reach a goal state from node n • \bad" heuristics can misguide search! Cost and Cost Estimation cont. Evaluation Function: f^(n) = g(n) + h^(n) Cost and Cost Estimation cont. • \True costs" of an optimal path from an initial state s to a final state: f (s). • For a node n on this path, f can be decomposed in the already performed steps with cost g(n) and the yet to perform steps with true cost h(n). • h^(n) can be an estimation which is greater or smaller than the true costs.
    [Show full text]
  • CAP 4630 Artificial Intelligence
    CAP 4630 Artificial Intelligence Instructor: Sam Ganzfried [email protected] 1 • http://www.ultimateaiclass.com/ • https://moodle.cis.fiu.edu/ • HW1 out 9/5 today, due 9/28 (maybe 10/3) – Remember that you have up to 4 late days to use throughout the semester. – https://www.cs.cmu.edu/~sganzfri/HW1_AI.pdf – http://ai.berkeley.edu/search.html • Office hours 2 • https://www.cs.cmu.edu/~sganzfri/Calendar_AI.docx • Midterm exam: on 10/19 • Final exam pushed back (likely on 12/12 instead of 12/5) • Extra lecture on NLP on 11/16 • Will likely only cover 1-1.5 lectures on logic and on planning • Only 4 homework assignments instead of 5 3 Heuristic functions 4 8-puzzle • The object of the puzzle is to slide the tiles horizontally or vertically into the empty space until the configuration matches the goal configuration. • The average solution cost for a randomly generated 8-puzzle instance is about 22 steps. The branching factor is about 3 (when the empty tile is in the middle, four moves are possible; when it is in a corner, two; and when it is along an edge, three). This means that an exhaustive tree search to depth 22 would look at about 3^22 = 3.1*10^10 states. A graph search would cut this down by a factor of about 170,000 because only 181,440 distinct states are reachable (homework exercise). This is a manageable number, but for a 15-puzzle, it would be 10^13, so we will need a good heuristic function.
    [Show full text]
  • Guided Local Search
    GUIDED LOCAL SEARCH Christos Voudouris, Edward P. K. Tsang and Abdullah Alsheddy Abstract Combinatorial explosion problem is a well known phenomenon that pre- vents complete algorithms from solving many real-life combinatorial optimization problems. In many situations, heuristic search methods are needed. This chapter de- scribes the principles of Guided Local Search (GLS) and Fast Local Search (FLS) and surveys their applications. GLS is a penalty-based meta-heuristic algorithm that sits on top of other local search algorithms, with the aim to improve their efficiency and robustness. FLS is a way of reducing the size of the neighbourhood to improve the efficiency of local search. The chapter also provides guidance for implement- ing and using GLS and FLS. Four problems, representative of general application categories, are examined with detailed information provided on how to build a GLS- based method in each case. Key words: Heuristic Search, Meta-Heuristics, Penalty-based Methods, Guided Local Search, Tabu Search, Constraint Satisfaction. 1 INTRODUCTION Many practical problems are NP-hard in nature, which means complete, construc- tive search is unlikely to satisfy our computational demand. For example, suppose Christos Voudouris Group Chief Technology Office, BT plc, Orion Building, mlb1/pp12, Marthlesham Heath, Ipswich, IP5 3RE, United Kingdom, e-mail: [email protected] Edward P. K. Tsang Department of Computer Science, Univeristy of Essex, Wivenhoe Park, Colchester, CO4 3SQ, United Kingdom, e-mail: [email protected] Abdullah Alsheddy Department of Computer Science, Univeristy of Essex, Wivenhoe Park, Colchester, CO4 3SQ, United Kingdom, e-mail: [email protected] 1 2 Christos Voudouris, Edward P.
    [Show full text]
  • Evaluation of Emerging Metaheuristic Strategies on Opimal Transmission Pricing
    Evaluation of Emerging Metaheuristic Strategies on Opimal Transmission Pricing José L. Rueda, Senior Member, IEEE István Erlich, Senior Member, IEEE Institute of Electrical Power Systems Institute of Electrical Power Systems University Duisburg-Essen University Duisburg-Essen Duisburg, Germany Duisburg, Germany [email protected] [email protected] Abstract--This paper provides a comparative assessment of the In practice, there are several factors that can influence the capabilities of three metaheuristic algorithms for solving the adoption of a particular scheme of transmission pricing. Thus, problem of optimal transmission pricing, whose formulation is existing literature on definition of pricing mechanisms is vast. based on principle of equivalent bilateral exchanges. Among the Particularly, the principle of Equivalent Bilateral Exchange selected algorithms are Covariance Matrix Adaptation (EBE), which was originally proposed in [5], has proven to be Evolution Strategy (CMA-ES), Linearized Biogeography-based useful for pool system by providing suitable price signals Optimization (LBBO), and a novel swarm variant of the Mean- reflecting variability in the usage rates and charges across Variance Mapping Optimization (MVMO-SM). The IEEE 30 transmission network. In [6], an optimization problem was bus system is used to perform numerical comparisons on devised based on EBE in order enable exploration of multiple convergence speed, achieved optimum solutions, and computing solutions in deciding equivalent bilateral exchanges. In the effort. 2011 competition on testing evolutionary algorithms for real- Index Terms--Equivalent bilateral exchanges, evolutionary world optimization problems (CEC11), this problem was mechanism, metaheuristics, transmission pricing. solved by different metaheuristic algorithms, most of them constituting extended or hybridized variants of genetic I.
    [Show full text]
  • (Stochastic) Local Search Algorithms
    DM841 DISCRETE OPTIMIZATION Part 2 – Heuristics (Stochastic) Local Search Algorithms Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Local Search Algorithms Basic Algorithms Outline Local Search Revisited 1. Local Search Algorithms 2. Basic Algorithms 3. Local Search Revisited Components 2 Local Search Algorithms Basic Algorithms Outline Local Search Revisited 1. Local Search Algorithms 2. Basic Algorithms 3. Local Search Revisited Components 3 Local Search Algorithms Basic Algorithms Local Search Algorithms Local Search Revisited Given a (combinatorial) optimization problem Π and one of its instances π: 1. search space S(π) I specified by the definition of (finite domain, integer) variables and their values handling implicit constraints I all together they determine the representation of candidate solutions I common solution representations are discrete structures such as: sequences, permutations, partitions, graphs (e.g., for SAT: array, sequence of truth assignments to propositional variables) Note: solution set S 0(π) ⊆ S(π) (e.g., for SAT: models of given formula) 4 Local Search Algorithms Basic Algorithms Local Search Algorithms (cntd) Local Search Revisited 2. evaluation function fπ : S(π) ! R I it handles the soft constraints and the objective function (e.g., for SAT: number of false clauses) S(π) 3. neighborhood function, Nπ : S ! 2 I defines for each solution s 2 S(π) a set of solutions N(s) ⊆ S(π) that are in some sense close to s. (e.g., for SAT: neighboring variable assignments differ in the truth value of exactly one variable) 5 Local Search Algorithms Basic Algorithms Local Search Algorithms (cntd) Local Search Revisited Further components [according to [HS]] 4.
    [Show full text]
  • Fast and Efficient Black Box Optimization Using the Parameter-Less Population Pyramid
    Fast and Efficient Black Box Optimization Using the Parameter-less Population Pyramid B. W. Goldman [email protected] Department of Computer Science and Engineering, Michigan State University, East Lansing, 48823, USA W. F. Punch [email protected] Department of Computer Science and Engineering, Michigan State University, East Lansing, 48823, USA doi:10.1162/EVCO_a_00148 Abstract The parameter-less population pyramid (P3) is a recently introduced method for per- forming evolutionary optimization without requiring any user-specified parameters. P3’s primary innovation is to replace the generational model with a pyramid of mul- tiple populations that are iteratively created and expanded. In combination with local search and advanced crossover, P3 scales to problem difficulty, exploiting previously learned information before adding more diversity. Across seven problems, each tested using on average 18 problem sizes, P3 outperformed all five advanced comparison algorithms. This improvement includes requiring fewer evaluations to find the global optimum and better fitness when using the same number of evaluations. Using both algorithm analysis and comparison, we find P3’s effectiveness is due to its ability to properly maintain, add, and exploit diversity. Unlike the best comparison algorithms, P3 was able to achieve this quality without any problem-specific tuning. Thus, unlike previous parameter-less methods, P3 does not sacrifice quality for applicability. There- fore we conclude that P3 is an efficient, general, parameter-less approach to black box optimization which is more effective than existing state-of-the-art techniques. Keywords Genetic algorithms, linkage learning, local search, parameter-less. 1 Introduction A primary purpose of evolutionary optimization is to efficiently find good solutions to challenging real-world problems with minimal prior knowledge about the problem itself.
    [Show full text]
  • Global Optimisation Through Hyper-Heuristics: Unfolding Population-Based Metaheuristics
    applied sciences Article Global Optimisation through Hyper-Heuristics: Unfolding Population-Based Metaheuristics Jorge M. Cruz-Duarte 1,† , José C. Ortiz-Bayliss 1,† , Ivan Amaya 1,*,† and Nelishia Pillay 2,† 1 School of Engineering and Sciences, Tecnologico de Monterrey, Av. Eugenio Garza Sada 2501 Sur, Monterrey 64849, NL, Mexico; [email protected] (J.M.C.-D.); [email protected] (J.C.O.-B.) 2 Department of Computer Science, University of Pretoria, Lynnwood Rd, Hatfield, Pretoria 0083, South Africa; [email protected] * Correspondence: [email protected]; Tel.: +52-(81)-8358-2000 † These authors contributed equally to this work. Abstract: Optimisation has been with us since before the first humans opened their eyes to natural phenomena that inspire technological progress. Nowadays, it is quite hard to find a solver from the overpopulation of metaheuristics that properly deals with a given problem. This is even considered an additional problem. In this work, we propose a heuristic-based solver model for continuous optimisation problems by extending the existing concepts present in the literature. We name such solvers ‘unfolded’ metaheuristics (uMHs) since they comprise a heterogeneous sequence of simple heuristics obtained from delegating the control operator in the standard metaheuristic scheme to a high-level strategy. Therefore, we tackle the Metaheuristic Composition Optimisation Problem by tailoring a particular uMH that deals with a specific application. We prove the feasibility of this model via a two-fold experiment employing several continuous optimisation problems and a collection of Citation: Cruz-Duarte, J.M.; diverse population-based operators with fixed dimensions from ten well-known metaheuristics in Ortiz-Bayliss, J.C.; Amaya, I.; Pillay, the literature.
    [Show full text]
  • A Multi-Objective Hyper-Heuristic Based on Choice Function
    A Multi-objective Hyper-heuristic based on Choice Function Mashael Maashi1, Ender Ozcan¨ 2, Graham Kendall3 1 2 3Automated Scheduling, Optimisation and Planning Research Group, University of Nottingham, Department of Computer Science, Jubilee Campus,Wollaton Road, Nottingham, NG8 1BB ,UK. 1email: [email protected] 2email: [email protected] 3The University of Nottingham Malaysia Campus email:[email protected] Abstract Hyper-heuristics are emerging methodologies that perform a search over the space of heuristics in an attempt to solve difficult computational opti- mization problems. We present a learning selection choice function based hyper-heuristic to solve multi-objective optimization problems. This high level approach controls and combines the strengths of three well-known multi-objective evolutionary algorithms (i.e. NSGAII, SPEA2 and MOGA), utilizing them as the low level heuristics. The performance of the pro- posed learning hyper-heuristic is investigated on the Walking Fish Group test suite which is a common benchmark for multi-objective optimization. Additionally, the proposed hyper-heuristic is applied to the vehicle crash- worthiness design problem as a real-world multi-objective problem. The experimental results demonstrate the effectiveness of the hyper-heuristic approach when compared to the performance of each low level heuristic run on its own, as well as being compared to other approaches including an adaptive multi-method search, namely AMALGAM. Keywords: Hyper-heuristic, metaheuristic, evolutionary algorithm, multi-objective optimization. 1. Introduction Most real-world problems are complex. Due to their (often) NP-hard nature, researchers and practitioners frequently resort to problem tailored heuristics to obtain a reasonable solution in a reasonable time.
    [Show full text]
  • On Metaheuristic Optimization Motivated by the Immune System
    Applied Mathematics, 2014, 5, 318-326 Published Online January 2014 (http://www.scirp.org/journal/am) http://dx.doi.org/10.4236/am.2014.52032 On Metaheuristic Optimization Motivated by the Immune System Mohammed Fathy Elettreby1,2*, Elsayd Ahmed2, Houari Boumedien Khenous1 1Department of Mathematics, Faculty of Science, King Khalid University, Abha, KSA 2Department of Mathematics, Faculty of Science, Mansoura University, Mansoura, Egypt Email: *[email protected] Received November 16, 2013; revised December 16, 2013; accepted December 23, 2013 Copyright © 2014 Mohammed Fathy Elettreby et al. This is an open access article distributed under the Creative Commons Attribu- tion License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. In accordance of the Creative Commons Attribution License all Copyrights © 2014 are reserved for SCIRP and the owner of the intellectual property Mohammed Fathy Elettreby et al. All Copyright © 2014 are guarded by law and by SCIRP as a guardian. ABSTRACT In this paper, we modify the general-purpose heuristic method called extremal optimization. We compare our results with the results of Boettcher and Percus [1]. Then, some multiobjective optimization problems are solved by using methods motivated by the immune system. KEYWORDS Multiobjective Optimization; Extremal Optimization; Immunememory and Metaheuristic 1. Introduction Multiobjective optimization problems (MOPs) [2] are existing in many situations in nature. Most realistic opti- mization problems require the simultaneous optimization of more than one objective function. In this case, it is unlikely that the different objectives would be optimized by the same alternative parameter choices. Hence, some trade-off between the criteria is needed to ensure a satisfactory problem.
    [Show full text]
  • Metaheuristics ``In the Large''
    Metaheuristics “In the Large” Jerry Swan∗, Steven Adriaensen, Alexander E. I. Brownlee, Kevin Hammond, Colin G. Johnson, Ahmed Kheiri, Faustyna Krawiec, J. J. Merelo, Leandro L. Minku, Ender Ozcan,¨ Gisele L. Pappa, Pablo Garc´ıa-S´anchez, Kenneth S¨orensen, Stefan Voß, Markus Wagner, David R. White Abstract Following decades of sustained improvement, metaheuristics are one of the great success stories of optimization research. However, in order for research in metaheuristics to avoid fragmentation and a lack of reproducibility, there is a pressing need for stronger scientific and computational infrastructure to sup- port the development, analysis and comparison of new approaches. To this end, we present the vision and progress of the “Metaheuristics ‘In the Large’ ” project. The conceptual uderpinnings of the project are: truly extensible algo- rithm templates that support reuse without modification, white box problem descriptions that provide generic support for the injection of domain specific knowledge, and remotely accessible frameworks, components and problems that will enhance reproducibility and accelerate the field’s progress. We ar- gue that, via principled choice of infrastructure support, the field can pur- sue a higher level of scientific enquiry. We describe our vision and report on progress, showing how the adoption of common protocols for all metaheuris- tics can help liberate the potential of the field, easing the exploration of the design space of metaheuristics. Keywords: Evolutionary Computation, Operational Research, Heuristic design, Heuristic methods, Architecture, Frameworks, Interoperability 1. Introduction arXiv:2011.09821v4 [cs.NE] 3 Jun 2021 Optimization problems have myriad real world applications [42] and have motivated a wealth of research since before the advent of the digital computer [25].
    [Show full text]