Evolutionary Computation Tutorial: Multi-Objective Optimisation Implementation Using Java

Total Page:16

File Type:pdf, Size:1020Kb

Evolutionary Computation Tutorial: Multi-Objective Optimisation Implementation Using Java Evolutionary Computation Evolutionary Computation Tutorial: Multi-objective optimisation implementation using Java Shan He School for Computational Science University of Birmingham Modules 02411 and 22313: EC Evolutionary Computation Outline Outline of Topics Existing EA/MOEA libraries and toolboxes Java MOEA framework How to Evaluate MOEAs? Notable Companies used Evolutionary Computation Conclusion Evolutionary Computation Existing EA/MOEA libraries and toolboxes Existing EA/MOEA frameworks and toolboxes I Matlab I Matworks Global Optimisation Toolbox I Genetic Algorithm Optimization Toolbox (GAOT) I Java I ECJ: Java Evolutionary Computation Toolkit I Watchmaker Framework for Evolutionary Computation I JGAP: Java Genetic Algorithm Package I MOEA framework I C/C++ I ParadisEO I Evolving Objects (EO): an Evolutionary Computation Framework I Open BEAGLE I Python I PyGMO I DEAP I inspyred: Bio-inspired Algorithms in Python Evolutionary Computation Java MOEA framework What is MOEA framework I \A free and open source Java library for developing and experimenting with multiobjective evolutionary algorithms (MOEAs) and other general-purpose multiobjective optimization algorithms" from MOEA framework web page I Many state-of-the-art multiobjective evolutionary algorithms are included: I Genetic algorithms: NSGA-II I Differential evolution I Particle swarm optimization I Genetic programming and Grammatical evolution I Other features: I Easy to use: minimum effor to define a problem, a algorithm and a experiment I Modular design: constructing new optimization algorithms from existing components I Very good documentation Evolutionary Computation Java MOEA framework How to use MOEA? I Download the compiled binaries from Sourceforge I Unzip and add libraries to Eclipse I To define a new problem, extend the AbstractProblem class I Use Executor, Instrumenter and Analyzer classes I Executor: class for constructing and executing runs of an algorithm I Instrumenter: class for analyzing the performance of algorithms I Analyzer: class for analysing the resulting Pareto approximation set and how it compares against a known reference set. Evolutionary Computation Java MOEA framework Example: multi-objective optimization in structural design Problem: design of a four bar plan truss with two objectives: I Minimise the volume of the truss I Minimise its joint displacement Evolutionary Computation Java MOEA framework Example: multi-objective optimization in structural design Mathematical formulation: ( p p f1(x) = L(2x1 + 2x2 + x3 + x4) min p p f (x) = FL ( 2 + 2 2 − 2 2 + 2 ) 2 E x1 x2 x3 x4 such that: p(F /σ) ≤ x1 ≤ 3(F /σ) 2(F /σ) ≤ x ≤ 3(F /σ) p 2 (1) 2(F /σ) ≤ x3 ≤ 3(F /σ) (F /σ) ≤ x4 ≤ 3(F /σ) where xi , i = 1; 2; 3; 4 are the cross sectional areas of the four bars, the acting force F = 10kN, Young's modulus of elasticity E = 2 × 105 kN/cm2, the length L = 200cm, the only nonzero stress component σ = 10kN/cm2. Evolutionary Computation How to Evaluate MOEAs? MOEAs Test Functions I ZDT benchmark: 6 problems, e.g., ZDT1 - ZDT6 I DTLZ benchmark: 7 problems, e.g., DTLZ1 - DTLZ7 I WFG benchmark: 9 problems, e.g., WFG1 - WFG9 I CEC2009 competition benchmark: 10 problems, e.g., UF1 - UF10 I One example: ZDT1: f1 = x1 q f = g · 1:0 − f1 2 g (2) 9 Pn g(x2; ··· ; xn) = 1:0 + n−1 i=2 xi 0 ≤ xi ≤ 1; i = 1; ··· ; n Evolutionary Computation How to Evaluate MOEAs? MOEAs Test Functions: ZDT1 Pareto Front of ZDT1: Evolutionary Computation How to Evaluate MOEAs? Multiobjective optimisation performance assessment I We need to design performance metrics or indicators to evaluate the algorithms' performance. I What characteristics should good Pareto optimal solutions have: I Minimize the distance of the Pareto front produced by the algorithm with respect to the true Pareto front if it is known I Maximize the spread of solutions found, so that we can have a distribution of vectors as smooth and uniform as possible. I Maximize the amount of elements of the Pareto optimal set found. I A myriad of ways of metrics, but can be classified as three kinds for measuring: I Convergence: converge to true Pareto front I Spread: maintain diversity of solution mainly in objective space I Both Convergence and Spread Evolutionary Computation How to Evaluate MOEAs? Multiobjective optimisation performance assessment I A few popular indicators I The hypervolume indicator: defines the size of multidimensional space that a set of solutions cover, without counting the same area twice. I Spread: calculates how uniformly spread across the objective space the solutions are within their set. I The cover function: calculates the proportion of one solution set that is dominated by another set. Thus, if C(X ; Y ) = 0:9, 90% of solutions in Y are dominated by solutions in X. I The Purity function: unifies the solutions returned by a set of algorithms on a particular problem, and returns the proportion of each that remain nondominated in the set. I A comprehensive indicators are listed in this paper: Yen, G.G. and He, Z.N. Performance Metric Ensemble for Multiobjective Evolutionary Algorithms, IEEE Trans on Evolutionary Computation, 18(1), Feb. 2014 (link to the paper) Evolutionary Computation How to Evaluate MOEAs? Example: the hypervolume indicator I Also called S metrics, essentially calculates how much of the objective space is dominated by a given nondominated set A I Formal definition: Let X be set of decision vectors for the considered problem and A = fx1; x2;; xtg 2 X a set of t decision vectors. The hypervolume indicator or S metrics S(A) gives the volume enclosed by the union of the polytopes p1; p2; ··· ; pt , where each pi is formed by the intersection of the following hyperplanes arising out of xi, along with the axes: for each axis in the objective space there exist a hyperplane perpendicular to the axis and passing through the point (f1(xi ); f2(xi ); ··· ; fk (xi )). Evolutionary Computation How to Evaluate MOEAs? Example: the hypervolume indicator Pareto Front of ZDT1: f1(x) Minimising f1(x) and f2(x) A B C PA D E PC PD PE f2(x) Evolutionary Computation Notable Companies used Evolutionary Computation Notable Companies used Evolutionary Computation I SolveIT Software: I Founded by Zbigniew Michalewicz, author of Genetic Algorithms + Data Structures I Uses Genetic Algorithm, PSO and other metaheurstic algorithms for advanced planning and scheduling optimization I Widely used in industries of wine, grain handling and mining such as Rio Tinto Group and BHP Billiton I 150 employees and 5 international offices I Acquired by Schneider Electric in 2012 but price was not divulged I Their Hybrid Optimisation System I Red Cedar Technology I DecisionVis Evolutionary Computation Conclusion Conclusion I Use library to solve your optimisation problems I Use performance metrics to evaluate your algorithm and solutions.
Recommended publications
  • Investigating the Parameter Space of Evolutionary Algorithms
    Sipper et al. BioData Mining (2018) 11:2 https://doi.org/10.1186/s13040-018-0164-x RESEARCH Open Access Investigating the parameter space of evolutionary algorithms Moshe Sipper1,2* ,WeixuanFu1,KarunaAhuja1 and Jason H. Moore1 *Correspondence: [email protected] Abstract 1Institute for Biomedical Informatics, Evolutionary computation (EC) has been widely applied to biological and biomedical University of Pennsylvania, data. The practice of EC involves the tuning of many parameters, such as population Philadelphia 19104-6021, PA, USA 2Department of Computer Science, size, generation count, selection size, and crossover and mutation rates. Through an Ben-Gurion University, Beer Sheva extensive series of experiments over multiple evolutionary algorithm implementations 8410501, Israel and 25 problems we show that parameter space tends to be rife with viable parameters, at least for the problems studied herein. We discuss the implications of this finding in practice for the researcher employing EC. Keywords: Evolutionary algorithms, Genetic programming, Meta-genetic algorithm, Parameter tuning, Hyper-parameter Introduction Evolutionary computation (EC) has been widely applied to biological and biomedical data [1–4]. One of the crucial tasks of the EC practitioner is the tuning of parameters. The fitness-select-vary paradigm comes with a plethora of parameters relating to the population, the generations, and the operators of selection, crossover, and mutation. It seems natural to ask whether the myriad parameters can be obtained through some clever methodology (perhaps even an evolutionary one) rather than by trial and error; indeed, as we shall see below, such methods have been previously devised. Our own interest in the issue of parameters stems partly from a desire to better understand evolutionary algo- rithms (EAs) and partly from our recent investigation into the design and implementation of an accessible artificial intelligence system [5].
    [Show full text]
  • Automatic Configuration of Multi-Objective Optimizers and Multi-Objective Configuration
    Chapter 1 Automatic Configuration of Multi-objective Optimizers and Multi-objective Configuration Leonardo C. T. Bezerra1, Manuel López-Ibáñez2, and Thomas Stützle3 Abstract Heuristic optimizers are an important tool in academia and industry, and their performance-optimizing configuration requires a significant amount of expertise. As the proper configuration of algorithms is a crucial aspect in the engineering of heuristic algorithms, a significant research effort has been dedicated over the last years towards moving this step to the computer and, thus, make it automatic. These research efforts go way beyond tuning only numerical parameters of already fully defined algorithms, but exploit automatic configuration as a means for automatic algorithm design. In this chapter, we review two main aspects where the research on automatic configuration and multi-objective optimization intersect. The first is the automatic configuration of multi-objective optimizers, where we discuss means and specific approaches. In addition, we detail a case study that shows how these approaches can be used to design new, high-performing multi-objective evolutionary algorithms. The second aspect is the research on multi-objective configuration, that is, the possibility of using multiple performance metrics for the evaluation of algorithm configurations. We highlight some few examples in this direction. Key words: multi-objective optimization, automatic algorithm configuration, automatic configuration of multi-objective optimizers, automatic configuration considering multiple objectives Instituto Metrópole Digital (IMD), Universidade Federal do Rio Grande do Norte (UFRN), Natal, RN, Brazil ([email protected]) · Alliance Manchester Business School, University of Manchester, UK ([email protected]) · IRIDIA, Université Libre de Bruxelles (ULB), Brussels, Belgium ([email protected]).
    [Show full text]
  • Application of Computational Intelligence to Explore and Analyze System Architecture and Design Alternatives
    Scholars' Mine Doctoral Dissertations Student Theses and Dissertations Fall 2019 Application of computational intelligence to explore and analyze system architecture and design alternatives Gene Lesinski Follow this and additional works at: https://scholarsmine.mst.edu/doctoral_dissertations Part of the Systems Engineering Commons Department: Engineering Management and Systems Engineering Recommended Citation Lesinski, Gene, "Application of computational intelligence to explore and analyze system architecture and design alternatives" (2019). Doctoral Dissertations. 2837. https://scholarsmine.mst.edu/doctoral_dissertations/2837 This thesis is brought to you by Scholars' Mine, a service of the Missouri S&T Library and Learning Resources. This work is protected by U. S. Copyright Law. Unauthorized use including reproduction for redistribution requires the permission of the copyright holder. For more information, please contact [email protected]. APPLICATION OF COMPUTATIONAL INTELLIGENCE TO EXPLORE AND ANALYZE SYSTEM ARCHITECTURE AND DESIGN ALTERNATIVES by EUGENE JOSEPH LESINSKI III A DISSERTATION Presented to the Faculty of the Graduate Faculty of the MISSOURI UNIVERSITY OF SCIENCE AND TECHNOLOGY In Partial Fulfillment of the Requirements for the Degree DOCTOR OF PHILOSOPHY in SYSTEMS ENGINEERING 2019 Approved by: Steven Corns, PhD, Advisor Suzanna Long, PhD Cihan Dagli, PhD Stephen Raper, PhD Donald Wunsch, PhD John Farr, PhD 2019 Eugene Joseph Lesinski III All Rights Reserved iii PUBLICATION DISSERTATION OPTION This dissertation consists of four peer reviewed publications that have been published, accepted, or that will be submitted, which have been formatted in the style used by the Missouri University of Science & Technology. Paper I: Pages 13-29 is a paper that has been published in IEEE Conference on Evolutionary Computation.
    [Show full text]
  • Pygad: an Intuitive Genetic Algorithm Python Library
    PyGAD: An Intuitive Genetic Algorithm Python Library Ahmed Fawzy Gad School of Electrical Engineering and Computer Science University of Ottawa Ottawa, ON, Canada [email protected] Abstract—This paper introduces PyGAD, an open-source easy- to-use Python library for building the genetic algorithm. PyGAD supports a wide range of parameters to give the user control over everything in its life cycle. This includes, but is not limited to, population, gene value range, gene data type, parent selection, crossover, and mutation. PyGAD is designed as a general- purpose optimization library that allows the user to customize the fitness function. Its usage consists of 3 main steps: build the fitness function, create an instance of the pygad.GA class, and calling the pygad.GA.run() method. The library supports training deep learning models created either with PyGAD itself or with frameworks like Keras and PyTorch. Given its stable state, PyGAD is also in active development to respond to the user’s requested features and enhancement received on GitHub github.com/ahmedfgad/GeneticAlgorithmPython. PyGAD comes with documentation pygad.readthedocs.io for further details and examples. Keywords— genetic algorithm, evolutionary algorithm, optimiza- tion, deep learning, Python, NumPy, Keras, PyTorch I. INTRODUCTION Nature has inspired computer scientists to create algorithms for solving computational problems. These naturally-inspired algorithms are called evolutionary algorithms (EAs) [1] where the initial solution (or individual) to a given problem is evolved across multiple iterations aiming to increase its quality. The EAs can be categorized by different factors like the number of solutions used. Some EAs evolve a single initial solution (e.g.
    [Show full text]
  • DEAP: Evolutionary Algorithms Made Easy
    Journal of Machine Learning Research 13 (2012) 2171-2175 Submitted 7/11; Revised 4/12; Published 7/12 DEAP: Evolutionary Algorithms Made Easy Felix-Antoine´ Fortin [email protected] Franc¸ois-Michel De Rainville [email protected] Marc-Andre´ Gardner [email protected] Marc Parizeau [email protected] Christian Gagne´ [email protected] Laboratoire de vision et systemes` numeriques´ Departement´ de genie´ electrique´ et de genie´ informatique Universite´ Laval Quebec´ (Quebec),´ Canada G1V 0A6 Editor: Mikio Braun Abstract DEAP is a novel evolutionary computation framework for rapid prototyping and testing of ideas. Its design departs from most other existing frameworks in that it seeks to make algorithms explicit and data structures transparent, as opposed to the more common black-box frameworks. Freely avail- able with extensive documentation at http://deap.gel.ulaval.ca, DEAP is an open source project under an LGPL license. Keywords: distributed evolutionary algorithms, software tools 1. Introduction Evolutionary Computation (EC) is a sophisticated field with very diverse techniques and mecha- nisms, where even well designed frameworks can become quite complicated under the hood. They thus strive to hide the implementation details as much as possible, by providing large libraries of high-level functionalities, often in many different flavours. This is the black-box software model (Roberts and Johnson, 1997). The more elaborate these boxes become, the more obscure they are, and the less likely the commoner is to ever take a peek under the hood to consider making changes. But using EC to solve real-world problems most often requires the customization of algorithms.
    [Show full text]
  • Evolutionary Algorithms and Other Metaheuristics in Water Resources: Current Status, Research Challenges and Future Directions*,**
    Environmental Modelling & Software 62 (2014) 271e299 Contents lists available at ScienceDirect Environmental Modelling & Software journal homepage: www.elsevier.com/locate/envsoft Position paper Evolutionary algorithms and other metaheuristics in water resources: Current status, research challenges and future directions*,** * H.R. Maier a, , Z. Kapelan b, J. Kasprzyk c, J. Kollat d, L.S. Matott e, M.C. Cunha f, G.C. Dandy a, M.S. Gibbs a, E. Keedwell b, A. Marchi a, A. Ostfeld g, D. Savic b, D.P. Solomatine h, i, J.A. Vrugt j, k, l, A.C. Zecchin a, B.S. Minsker m, E.J. Barbour n, G. Kuczera o, F. Pasha p, A. Castelletti q, M. Giuliani q, P.M. Reed r a School of Civil, Environmental and Mining Engineering, The University of Adelaide, Adelaide, SA 5005, Australia b College of Engineering, Mathematics and Physical Sciences, University of Exeter, Harrison Building, North Park Road, Exeter EX4 4QF, United Kingdom c Civil Environmental and Architectural Engineering Department, University of Colorado Boulder, UCB 428, Boulder, CO 80309, USA d DecisionVis, 315 S. Allen St., Ste 321, State College, PA 16801, USA e Center for Computational Research, University at Buffalo, Buffalo, NY 14203, USA f Department of Civil Engineering, University of Coimbra, Rua Luís Reis Santos e Polo II, 3030-788 Coimbra, Portugal g Civil and Environmental Engineering, Technion e Israel Institute of Technology, Haifa 32000, Israel h UNESCO-IHE Institute for Water Education, Delft, The Netherlands i Water Resources Section, Delft University of Technology, The Netherlands j Department of Civil and Environmental Engineering, University of California Irvine, Irvine, CA, USA k Department of Earth System Science, University of California Irvine, Irvine, CA, USA l Institute for Biodiversity and Ecosystem Dynamics, University of Amsterdam, Amsterdam, The Netherlands m Department of Civil and Environmental Engineering, 205 N.
    [Show full text]
  • An Open-Source Nature-Inspired Optimization Framework in Python
    EvoloPy: An Open-source Nature-inspired Optimization Framework in Python Hossam Faris1, Ibrahim Aljarah1, Seyedali Mirjalili2, Pedro A. Castillo3 and Juan J. Merelo3 1Department of Business Information Technology, King Abdullah II School for Information Technology, The University of Jordan, Amman, Jordan 2School of Information and Communication Technology, Griffith University, Nathan, Brisbane, QLD 4111, Australia 3ETSIIT-CITIC, University of Granada, Granada, Spain Keywords: Evolutionary, Swarm Optimization, Metaheuristic, Optimization, Python, Framework. Abstract: EvoloPy is an open source and cross-platform Python framework that implements a wide range of classical and recent nature-inspired metaheuristic algorithms. The goal of this framework is to facilitate the use of metaheuristic algorithms by non-specialists coming from different domains. With a simple interface and minimal dependencies, it is easier for researchers and practitioners to utilize EvoloPy for optimizing and benchmarking their own defined problems using the most powerful metaheuristic optimizers in the literature. This framework facilitates designing new algorithms or improving, hybridizing and analyzing the current ones. The source code of EvoloPy is publicly available at GitHub (https://github.com/7ossam81/EvoloPy). 1 INTRODUCTION between the swarm’s individuals. On the other hand, evolutionary-based algorithms In general, nature-inspired algorithms are population- are inspired by some concepts from the Darwinian based metaheuristics which are inspired by different theory about evolution and natural selection. Such phenomena in nature. Due to their stochastic and non- algorithms include Genetic algorithm (GA) (Hol- deterministic nature, there has been a growing interest land, 1992), Genetic Programming (GP)(Koza, 1992), in investigating their applications for solving complex and Evolution Strategy (ES) (Beyer and Schwefel, problems when the search space is extremely large 2002).
    [Show full text]
  • Arxiv:1903.02915V2 [Cs.NE] 17 Apr 2019 an Important Factor That Has Ignited the Widespread Adop- Namic Problems
    jMetalPy: a Python Framework for Multi-Objective Optimization with Metaheuristics Antonio Ben´ıtez-Hidalgoa, Antonio J. Nebroa, Jose´ Garc´ıa-Nietoa, Izaskun Oregib, Javier Del Serb,c,d aDepartamento de Lenguajes y Ciencias de la Computaci´on,Ada Byron Research Building, University of M´alaga, 29071 M´alaga, Spain bTECNALIA, 48160 Derio, Spain cUniversity of the Basque Country (UPV/EHU), 48013 Bilbao, Spain dBasque Center for Applied Mathematics (BCAM), 48009 Bilbao, Spain Abstract This paper describes jMetalPy, an object-oriented Python-based framework for multi-objective optimization with metaheuristic techniques. Building upon our experiences with the well-known jMetal framework, we have developed a new multi-objective optimization software platform aiming not only at replicating the former one in a different programming language, but also at taking advantage of the full feature set of Python, including its facilities for fast prototyping and the large amount of available libraries for data processing, data analysis, data visualization, and high-performance computing. As a result, jMetalPy provides an environment for solving multi-objective optimization problems focused not only on traditional metaheuristics, but also on techniques supporting preference articulation and dynamic problems, along with a rich set of features related to the automatic generation of statistical data from the results generated, as well as the real-time and interactive visualization of the Pareto front approximations produced by the algorithms. jMetalPy offers additionally support for parallel computing in multicore and cluster systems. We include some use cases to explore the main features of jMetalPy and to illustrate how to work with it. Keywords: Multi-Objective Optimization, Metaheuristics, Software Framework, Python, Statistical Analysis, Visualization 1.
    [Show full text]
  • And Many-Objective Evolutionary Algorithms
    Automatically Designing State-of-the-Art Multi- and Many-Objective Evolutionary Algorithms Leonardo C. T. Bezerra [email protected] Instituto Metrópole Digital (IMD), Universidade Federal do Rio Grande do Norte, Natal, RN, Brazil Downloaded from http://direct.mit.edu/evco/article-pdf/28/2/195/1858883/evco_a_00263.pdf by guest on 28 September 2021 Manuel López-Ibáñez [email protected] Alliance Manchester Business School, University of Manchester, UK Thomas Stützle [email protected] IRIDIA, CoDE, Université Libre de Bruxelles, Belgium https://doi.org/10.1162/evco_a_00263 Abstract A recent comparison of well-established multiobjective evolutionary algorithms (MOEAs) has helped better identify the current state-of-the-art by considering (i) pa- rameter tuning through automatic configuration, (ii) a wide range of different setups, and (iii) various performance metrics. Here, we automatically devise MOEAs with ver- ified state-of-the-art performance for multi- and many-objective continuous optimiza- tion. Our work is based on two main considerations. The first is that high-performing algorithms can be obtained from a configurable algorithmic framework in an auto- mated way. The second is that multiple performance metrics may be required to guide this automatic design process. In the first part of this work, we extend our previously proposed algorithmic framework, increasing the number of MOEAs, underlying evo- lutionary algorithms, and search paradigms that it comprises. These components can be combined following a general MOEA template, and an automatic configuration method is used to instantiate high-performing MOEA designs that optimize a given performance metric and present state-of-the-art performance.
    [Show full text]
  • Optimal Solution for Santa Fe Trail Ant Problem Using MOEA Framework
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395 -0056 Volume: 04 Issue: 04 | Apr -2017 www.irjet.net p-ISSN: 2395-0072 Optimal Solution for Santa Fe Trail Ant Problem using MOEA Framework Parneet Aulakh1, Sarvesh Chopra2 1Research Scholar, CT institute of engineering management and technology, Jalandhar,Punjab,India 2 Assistant Professor, CT institute of engineering management and technology, Jalandhar, India Abstract : Multi-objective optimization is also a most real problems. Pareto optimality considers significant topic for research from past few decades. This solutions to be superior or inferior to another solution is because of the multi-objective nature of real world only when it is superior in all objectives or inferior in all problems. Most of the real world problems are complex objectives, respectively. The Santa Fe Ant Problem is a and versatile in nature, and quite often need more than conventional ideal problem that has been studied over one conflicting objective functions to be optimized the past two decades and is still being rigorously simultaneously. Researchers have developed many multi- researched. Santa Fe Trail is well-known of its objective optimization procedures. Santa Fe Trail Ant reputation of being hard due to evolutionary computing Problem has been widely used to analyze experiment and methods not solving it at much higher effectiveness than examine various evolutionary computing methods and random search. NSGA II is an evolutionary algorithm in multi-objective optimization problems. In this, Santa Fe which the main advantage is that it handles multi- Trail Ant problem is solved in its best optimization valued objective solution given in sets of solutions, which solution using various techniques includes mainly Non- provide computation of an approximation of the entire dominated Sorting Genetic Algorithm (NSGA II), Simulated Pareto front.
    [Show full text]
  • Analysis of Open-Source Hyperparameter Optimization Software Trends
    International Journal of Advanced Culture Technology Vol.7 No.4 56-62 (2019) DOI 10.17703/IJACT.2019.7.4.56 IJACT 19-12-7 Analysis of Open-Source Hyperparameter Optimization Software Trends Yo-Seob Lee1, Phil-Joo Moon2* 1Professor, School of ICT Convergence, Pyeongtaek University E-mail: [email protected] 2Professor, Dept. of Information & Communication, Pyeongtaek University E-mail: [email protected] Abstract Recently, research using artificial neural networks has further expanded the field of neural network optimization and automatic structuring from improving inference accuracy. The performance of the machine learning algorithm depends on how the hyperparameters are configured. Open-source hyperparameter optimization software can be an important step forward in improving the performance of machine learning algorithms. In this paper, we review open-source hyperparameter optimization softwares. Keywords: Hyperparameter Optimization, Machine Learning, Deep Learning, AutoML 1. INTRODUCTION Recently, researches using artificial neural networks have further expanded the field of neural network optimization and automatic structuring from improving the inference accuracy of image, video and natural language based tasks. In addition, with recent interest in complex and computationally expensive machine learning models with many hyperparameters, such as the Automatic Machine Learning(AutoML) framework and deep neural networks, research on hyperparameter optimization(HPO) has begun again. The performance of machine learning algorithms significantly
    [Show full text]
  • Automatically Designing State-Of-The-Art Multi- and Many-Objective Evolutionary Algorithms
    Automatically Designing State-of-the-Art Multi- and Many-Objective Evolutionary Algorithms Leonardo C. T. Bezerra [email protected] Instituto Metrópole Digital (IMD), Universidade Federal do Rio Grande do Norte, Natal, RN, Brazil Manuel López-Ibáñez [email protected] Alliance Manchester Business School, University of Manchester, UK Thomas Stützle [email protected] IRIDIA, CoDE, Université Libre de Bruxelles, Belgium https://doi.org/10.1162/evco_a_00263 Abstract A recent comparison of well-established multiobjective evolutionary algorithms (MOEAs) has helped better identify the current state-of-the-art by considering (i) pa- rameter tuning through automatic configuration, (ii) a wide range of different setups, and (iii) various performance metrics. Here, we automatically devise MOEAs with ver- ified state-of-the-art performance for multi- and many-objective continuous optimiza- tion. Our work is based on two main considerations. The first is that high-performing algorithms can be obtained from a configurable algorithmic framework in an auto- mated way. The second is that multiple performance metrics may be required to guide this automatic design process. In the first part of this work, we extend our previously proposed algorithmic framework, increasing the number of MOEAs, underlying evo- lutionary algorithms, and search paradigms that it comprises. These components can be combined following a general MOEA template, and an automatic configuration method is used to instantiate high-performing MOEA designs that optimize a given performance metric and present state-of-the-art performance. In the second part, we propose a multiobjective formulation for the automatic MOEA design, which proves critical for the context of many-objective optimization due to the disagreement of es- tablished performance metrics.
    [Show full text]