Genetic Algorithms for Applied Path Planning a Thesis Presented in Partial Ful Llment of the Honors Program Vincent R

Total Page:16

File Type:pdf, Size:1020Kb

Genetic Algorithms for Applied Path Planning a Thesis Presented in Partial Ful Llment of the Honors Program Vincent R Genetic Algorithms for Applied Path Planning A Thesis Presented in Partial Ful llment of the Honors Program Vincent R. Ragusa Abstract Path planning is the computational task of choosing a path through an environment. As a task humans do hundreds of times a day, it may seem that path planning is an easy task, and perhaps naturally suited for a computer to solve. This is not the case however. There are many ways in which NP-Hard problems like path planning can be made easier for computers to solve, but the most signi cant of these is the use of approximation algorithms. One such approximation algorithm is called a genetic algorithm. Genetic algorithms belong to a an area of computer science called evolutionary computation. The techniques used in evolutionary computation algorithms are modeled after the principles of Darwinian evolution by natural selection. Solutions to the problem are literally bred for their problem solving ability through many generations of selective breeding. The goal of the research presented is to examine the viability of genetic algorithms as a practical solution to the path planning problem. Various modi cations to a well known genetic algorithm (NSGA-II) were implemented and tested experimentally to determine if the modi cation had an e ect on the operational eciency of the algorithm. Two new forms of crossover were implemented with positive results. The notion of mass extinction driving evolution was tested with inconclusive results. A path correction algorithm called make valid was created which has proven to be extremely powerful. Finally several additional objective functions were tested including a path smoothness measure and an obstacle intrusion measure, the latter showing an enormous positive result. Department of Computer Science Under the supervision of Dr. H. David Mathias Florida Southern College May 2017 Contents Abstract Acknowledgments 1 Introduction 1 1.1 Optimization Problems . 1 1.1.1 Single-Objective Optimization . 1 1.1.2 Multi-Objective Optimization . 2 1.1.3 Computational Complexity . 3 1.2 Path Planning . 6 1.3 UAVs and MAVs . 7 1.4 Genetic Algorithms . 7 1.4.1 Solution Encoding Scheme . 8 1.4.2 Genetic Operators . 8 1.4.3 Selection and Niching . 10 1.4.4 Tuning a Genetic Algorithm . 13 2 Problem Statement 13 3 The Algorithm 14 3.1 Path nder . 14 3.1.1 Genome & Solution Encoding . 14 3.1.2 Objectives . 15 3.1.3 Algorithm Structure . 15 3.1.4 Genetic Operators . 16 4 Experiments and Results 16 4.1 Crossover and Mass Extinction . 17 4.1.1 Tested Components . 17 4.1.2 Experimental Method . 18 4.1.3 Results . 20 4.2 Crossover, Obstacle Intrusion, Path Correction, and Mutation Size . 22 4.2.1 Tested Components . 22 4.2.2 Experimental Method . 24 4.2.3 Results . 26 5 Final Conclusions 30 5.1 Crossover and Mass Extinction . 30 5.2 Crossover, Obstacle Intrusion, and Path Correction . 31 References Acknowledgments I would like to thank Florida Southern College for nancial and other support for this re- search, 3DR for providing educational pricing and support of open source projects important to research, commercial, and hobbyist projects for MAVs, and Peter Barker for assistance with the drone ight simulator. I would also like to thank Annie Wu and Vera Kazakova for valuable feedback and collaboration, Isabel Loyd and Susan Serrano for assistance with the statistical analysis, and David Mathias for all of his support and guidance. 1 Introduction The very idea of eciency is born from asking "Why is this task accomplished in this way?" and following that question with "Is there a better way?". This way of arriving at eciency leaves a lot of room to interpret "better" and as such the task of nding "the best" of anything can be extremely complicated. In mathematics and in computer science the task of nding the best of something is called an optimization problem. 1.1 Optimization Problems Optimization problems can be thought of as search problems where the task is to search for the optimal solution to a problem from among a pool of candidate solutions. There are typically many candidate solutions, each, usually, with an associated value (or values) which are desired to either be maxamized or minimized[1]. It is convenient to call all optimization problems minimization problems because any value you wish to maximize can be multiplied by ( 1) and then minimized to achieve the same e ect. Also note that it is common for there to be more than one candidate solution which achieves the optimal value[1]. 1.1.1 Single-Objective Optimization Single-objective optimization problems seek the minimum of a single value. Single-objective problems are typically solved by mathematical methods such as evaluating derivatives and gradient descent as well as basic computational methods like greedy strategies. For example, consider a single-objective optimization problem where the goal is to minimize the potential energy of a spring U(x). 1 U(x) = kx2 2 Evaluating dU = 0, dx d U(x) = kx = 0 dx ) x = 0 it is seen that there is a critical point when x = 0. Furthermore evaluating d2U at the critical dx2 point, d2U = k 2 dx x=0 k > 0 reveals that the critical point is a minimum because the second derivative is positive at the critical point. Therefore the minimum of U(x) is at the point x = 0, or when the spring is neither compressed or stretched. Mathematical methods are often the easiest method for solving single objective optimization problems. 1 1.1.2 Multi-Objective Optimization Multi-objective optimization problems seek to minimize n values simultaneously. Under ideal circumstances each objective is independent and the problem reduces to solving n single-objective optimization problems in parallel. This is almost never the case however as a single property of a candidate solution usually a ects multiple objective values. Consider the task of maximizing the volume of a box (minimizing the volume multiplied by ( 1)) and minimizing the surface area of the box. Note x; y; z 2 (0; 1). V (x; y; z) = xyz A(x; y; z) = 2(xy + xz + yz) If an attempt is made to minimize V (x; y; z) using the same mathematical method used for the spring example in section 1.1.1: rV = (yz; xz; xy) = (0; 0; 0) ) (x = y = 0) OR (x = z = 0) OR (y = z = 0) a contradiction is reached because x, y, and z cannot equal zero. A second attempt, this time to minimize A(x; y; z) reveals: rA = 2(y + z; x + z; x + y) = (0; 0; 0) ) (y + z = 0) AND (x + z = 0) AND (x + y = 0) Solving by elimination: (x; y; z) = (0; 0; 0) Again a contradiction is reached because x, y, and z cannot equal zero. Finally, let R(x; y; z) be de ned as the ratio of the area to volume: A(x; y; z) 2(xy + xz + yz) 2 2 2 R(x; y; z) = = = + + V (x; y; z) xyz z y x Searching for critical points reveals: 1 1 1 rR = 2( ; ; ) = (0; 0; 0) z2 y2 x2 which has no solutions. It seems that without some new mathematical or computational tools these problems cannot be solved. Scalarization Sometimes it is still useful to attempt to optimize a multi-objective problem by using a method known as scalarization. Scalarization is simply the transformation of n objective functions ff1; ¡¡¡ ; fng into a single function F via a weighted sum[2], where ci 6= 0 is the weight on fi. Xn F (p) = cifi(p) i=1 2 Often, the weights sum to 1 so that each weight represents a percentage of the nal sum. Xn ci = 1 i=1 Individual objectives can be given preference by giving them larger weights. Raising the objective function to a di erent power k 6= 0 can be used to emphasize (k > 0) or de- emphasize (k < 0) changes in an objective function. Xn ki F (p) = ci(fi(p)) i=1 The biggest drawback of scalarization is that contributions by individual tness measures are anonymous, hiding useful data about the search space[2]. Additionally, it falls victim to things like change blindness, or when multiple tness functions change but the weighted sum has a net change of zero. Scalarization is often abandoned for better methods of dealing with multiple objectives simultaneously. Pareto Optimization One popular method of handling multiple objectives in an optimization problem is to rank them not via the tness values directly, but by their relative Pareto eciency. Simply put, a candidate solution to a multi-objective optimization problem is Pareto ecient (or Pareto optimal) if it cannot make any progress in one of its objectives without causing negative changes in some or all of the other objectives[3]. This is more easily illustrated by rst de ning a vector comparison operator "". Deb et al[4] de ne this operator such that vector v1 is partially less than vector v2 when: v1 v2 (8i j v2;i v1;i) ^ (9i j v1;i < v2;i) With this, comparing two solutions is done by recognizing each objective's tness score as a component of a tness vector with dimension equal to the number of objectives. If the tness vector of one solution is partially less than that of another then the former dominates the latter. With this new operator Pareto optimal can be de ned as a solution that cannot be dominated by another solution. Note that there can be (and often are) an in nite number of Pareto optimal solutions to a given multi-objective optimization problem.
Recommended publications
  • Completely Derandomized Self-Adaptation in Evolution Strategies
    Completely Derandomized Self-Adaptation in Evolution Strategies Nikolaus Hansen [email protected] Technische Universitat¨ Berlin, Fachgebiet fur¨ Bionik, Sekr. ACK 1, Ackerstr. 71–76, 13355 Berlin, Germany Andreas Ostermeier [email protected] Technische Universitat¨ Berlin, Fachgebiet fur¨ Bionik, Sekr. ACK 1, Ackerstr. 71–76, 13355 Berlin, Germany Abstract This paper puts forward two useful methods for self-adaptation of the mutation dis- tribution – the concepts of derandomization and cumulation. Principle shortcomings of the concept of mutative strategy parameter control and two levels of derandomization are reviewed. Basic demands on the self-adaptation of arbitrary (normal) mutation dis- tributions are developed. Applying arbitrary, normal mutation distributions is equiv- alent to applying a general, linear problem encoding. The underlying objective of mutative strategy parameter control is roughly to favor previously selected mutation steps in the future. If this objective is pursued rigor- ously, a completely derandomized self-adaptation scheme results, which adapts arbi- trary normal mutation distributions. This scheme, called covariance matrix adaptation (CMA), meets the previously stated demands. It can still be considerably improved by cumulation – utilizing an evolution path rather than single search steps. Simulations on various test functions reveal local and global search properties of the evolution strategy with and without covariance matrix adaptation. Their per- formances are comparable only on perfectly scaled functions. On badly scaled, non- separable functions usually a speed up factor of several orders of magnitude is ob- served. On moderately mis-scaled functions a speed up factor of three to ten can be expected. Keywords Evolution strategy, self-adaptation, strategy parameter control, step size control, de- randomization, derandomized self-adaptation, covariance matrix adaptation, evolu- tion path, cumulation, cumulative path length control.
    [Show full text]
  • A New Biogeography-Based Optimization Approach Based on Shannon-Wiener Diversity Index to Pid Tuning in Multivariable System
    ABCM Symposium Series in Mechatronics - Vol. 5 Section III – Emerging Technologies and AI Applications Copyright © 2012 by ABCM Page 592 A NEW BIOGEOGRAPHY-BASED OPTIMIZATION APPROACH BASED ON SHANNON-WIENER DIVERSITY INDEX TO PID TUNING IN MULTIVARIABLE SYSTEM Marsil de Athayde Costa e Silva, [email protected] Undergraduate course of Mechatronics Engineering Camila da Costa Silveira, [email protected] Leandro dos Santos Coelho, [email protected] Industrial and Systems Engineering Graduate Program, PPGEPS Pontifical Catholic University of Parana, PUCPR Imaculada Conceição, 1155, Zip code 80215-901, Curitiba, Parana, Brazil Abstract. Proportional-integral-derivative (PID) control is the most popular control architecture used in industrial problems. Many techniques have been proposed to tune the gains for the PID controller. Over the last few years, as an alternative to the conventional mathematical approaches, modern metaheuristics, such as evolutionary computation and swarm intelligence paradigms, have been given much attention by many researchers due to their ability to find good solutions in PID tuning. As a modern metaheuristic method, Biogeography-based optimization (BBO) is a generalization of biogeography to evolutionary algorithm inspired on the mathematical model of organism distribution in biological systems. BBO is an evolutionary process that achieves information sharing by biogeography-based migration operators. This paper proposes a modification for the BBO using a diversity index, called Shannon-wiener index (SW-BBO), for tune the gains of the PID controller in a multivariable system. Results show that the proposed SW-BBO approach is efficient to obtain high quality solutions in PID tuning. Keywords: PID control, biogeography-based optimization, Shannon-Wiener index, multivariable systems.
    [Show full text]
  • The Roles of Evolutionary Computation, Fitness Landscape, Constructive Methods and Local Searches in the Development of Adaptive Systems for Infrastructure Planning
    International Symposium for Next Generation Infrastructure October 1-4, 2013, Wollongong, Australia The roles of evolutionary computation, fitness landscape, constructive methods and local searches in the development of adaptive systems for infrastructure planning Mehrdad Amirghasemi a* Reza Zamani a Abstract: Modelling and systems simulation for improving city planning, and traffic equilibrium are involved with the development of adaptive systems that can learn and respond to the environment intelligently. By employing sophisticated techniques which highly support infrastructure planning and design, evolutionary computation can play a key role in the development of such systems. The key to presenting solution strategies for these systems is fitness landscape which makes some problems hard and some problems easy to tackle. Moreover, constructive methods and local searches can assist evolutionary searches to improve their performance. In this paper, in the context of infrastructure, in general, and city planning, and traffic equilibrium, in particular, the integration of these four concepts is briefly discussed. Key words: Evolutionary computation; Local search; Infrastructure; City planning; Traffic equilibrium. I. Introduction Infrastructure is a term used to indicate both organizational and physical structures a society or an enterprise needs to function. In effect, all facilities and services needed for the operation of an economy is gathered under the umbrella term of “Infrastructure”. Each Infrastructure includes connected elements which are structurally related and each element can affect the other elements. It is through this interconnection that these elements collectively provide a basis for structural development of a society or an enterprise. Viewed in an operative level, infrastructure makes the production of goods and services required in the society more straightforward and the distribution of finished products to market easier.
    [Show full text]
  • Sustainable Growth and Synchronization in Protocell Models
    life Article Sustainable Growth and Synchronization in Protocell Models Roberto Serra 1,2,3 and Marco Villani 1,2,* 1 Department of Physics, Informatics and Mathematics, Modena and Reggio Emilia University, Via Campi 213/A, 41125 Modena, Italy 2 European Centre for Living Technology, Ca’ Bottacin, Dorsoduro 3911, Calle Crosera, 30123 Venice, Italy 3 Institute for Advanced Study, University of Amsterdam, Oude Turfmarkt 147, 1012 GC Amsterdam, The Netherlands * Correspondence: [email protected] Received: 19 July 2019; Accepted: 14 August 2019; Published: 21 August 2019 Abstract: The growth of a population of protocells requires that the two key processes of replication of the protogenetic material and reproduction of the whole protocell take place at the same rate. While in many ODE-based models such synchronization spontaneously develops, this does not happen in the important case of quadratic growth terms. Here we show that spontaneous synchronization can be recovered (i) by requiring that the transmembrane diffusion of precursors takes place at a finite rate, or (ii) by introducing a finite lifetime of the molecular complexes. We then consider reaction networks that grow by the addition of newly synthesized chemicals in a binary polymer model, and analyze their behaviors in growing and dividing protocells, thereby confirming the importance of (i) and (ii) for synchronization. We describe some interesting phenomena (like long-term oscillations of duplication times) and show that the presence of food-generated autocatalytic cycles is not sufficient to guarantee synchronization: in the case of cycles with a complex structure, it is often observed that only some subcycles survive and synchronize, while others die out.
    [Show full text]
  • Arxiv:1803.03453V4 [Cs.NE] 21 Nov 2019
    The Surprising Creativity of Digital Evolution: A Collection of Anecdotes from the Evolutionary Computation and Artificial Life Research Communities Joel Lehman1†, Jeff Clune1, 2†, Dusan Misevic3†, Christoph Adami4, Lee Altenberg5, Julie Beaulieu6, Peter J Bentley7, Samuel Bernard8, Guillaume Beslon9, David M Bryson4, Patryk Chrabaszcz11, Nick Cheney2, Antoine Cully12, Stephane Doncieux13, Fred C Dyer4, Kai Olav Ellefsen14, Robert Feldt15, Stephan Fischer16, Stephanie Forrest17, Antoine Fr´enoy18, Christian Gagn´e6 Leni Le Goff13, Laura M Grabowski19, Babak Hodjat20, Frank Hutter11, Laurent Keller21, Carole Knibbe9, Peter Krcah22, Richard E Lenski4, Hod Lipson23, Robert MacCurdy24, Carlos Maestre13, Risto Miikkulainen26, Sara Mitri21, David E Moriarty27, Jean-Baptiste Mouret28, Anh Nguyen2, Charles Ofria4, Marc Parizeau 6, David Parsons9, Robert T Pennock4, William F Punch4, Thomas S Ray29, Marc Schoenauer30, Eric Schulte17, Karl Sims, Kenneth O Stanley1,31, Fran¸coisTaddei3, Danesh Tarapore32, Simon Thibault6, Westley Weimer33, Richard Watson34, Jason Yosinski1 †Organizing lead authors 1 Uber AI Labs, San Francisco, CA, USA 2 University of Wyoming, Laramie, WY, USA 3 Center for Research and Interdisciplinarity, Paris, France 4 Michigan State University, East Lansing, MI, USA 5 Univeristy of Hawai‘i at Manoa, HI, USA 6 Universit´eLaval, Quebec City, Quebec, Canada 7 University College London, London, UK 8 INRIA, Institut Camille Jordan, CNRS, UMR5208, 69622 Villeurbanne, France 9 Universit´ede Lyon, INRIA, CNRS, LIRIS UMR5205, INSA, UCBL,
    [Show full text]
  • A Note on Evolutionary Algorithms and Its Applications
    Bhargava, S. (2013). A Note on Evolutionary Algorithms and Its Applications. Adults Learning Mathematics: An International Journal, 8(1), 31-45 A Note on Evolutionary Algorithms and Its Applications Shifali Bhargava Dept. of Mathematics, B.S.A. College, Mathura (U.P)- India. <[email protected]> Abstract This paper introduces evolutionary algorithms with its applications in multi-objective optimization. Here elitist and non-elitist multiobjective evolutionary algorithms are discussed with their advantages and disadvantages. We also discuss constrained multiobjective evolutionary algorithms and their applications in various areas. Key words: evolutionary algorithms, multi-objective optimization, pareto-optimality, elitist. Introduction The term evolutionary algorithm (EA) stands for a class of stochastic optimization methods that simulate the process of natural evolution. The origins of EAs can be traced back to the late 1950s, and since the 1970’s several evolutionary methodologies have been proposed, mainly genetic algorithms, evolutionary programming, and evolution strategies. All of these approaches operate on a set of candidate solutions. Using strong simplifications, this set is subsequently modified by the two basic principles of evolution: selection and variation. Selection represents the competition for resources among living beings. Some are better than others and more likely to survive and to reproduce their genetic information. In evolutionary algorithms, natural selection is simulated by a stochastic selection process. Each solution is given a chance to reproduce a certain number of times, dependent on their quality. Thereby, quality is assessed by evaluating the individuals and assigning them scalar fitness values. The other principle, variation, imitates natural capability of creating “new” living beings by means of recombination and mutation.
    [Show full text]
  • Analysis of Evolutionary Algorithms
    Why We Do Not Evolve Software? Analysis of Evolutionary Algorithms Roman V. Yampolskiy Computer Engineering and Computer Science University of Louisville [email protected] Abstract In this paper, we review the state-of-the-art results in evolutionary computation and observe that we don’t evolve non-trivial software from scratch and with no human intervention. A number of possible explanations are considered, but we conclude that computational complexity of the problem prevents it from being solved as currently attempted. A detailed analysis of necessary and available computational resources is provided to support our findings. Keywords: Darwinian Algorithm, Genetic Algorithm, Genetic Programming, Optimization. “We point out a curious philosophical implication of the algorithmic perspective: if the origin of life is identified with the transition from trivial to non-trivial information processing – e.g. from something akin to a Turing machine capable of a single (or limited set of) computation(s) to a universal Turing machine capable of constructing any computable object (within a universality class) – then a precise point of transition from non-life to life may actually be undecidable in the logical sense. This would likely have very important philosophical implications, particularly in our interpretation of life as a predictable outcome of physical law.” [1]. 1. Introduction On April 1, 2016 Dr. Yampolskiy posted the following to his social media accounts: “Google just announced major layoffs of programmers. Future software development and updates will be done mostly via recursive self-improvement by evolving deep neural networks”. The joke got a number of “likes” but also, interestingly, a few requests from journalists for interviews on this “developing story”.
    [Show full text]
  • Genetic Algorithms and Evolutionary Computation
    126 IJCSNS International Journal of Computer Science and Network Security, VOL.10 No.12, December 2010 Genetic Algorithms and Evolutionary Computation Ms.Nagori Meghna, Faculty of EngineeringGovernment EnggCollege Aurangabad, -431432, India Ms. KaleJyoti Faculty of Engineering MGM’c Engg College Nanded, -411402, India Abstract from the wild type), and dairy cows (with immense udders Generally speaking, genetic algorithms are simulations of far larger than would be required just for nourishing evolution, of what kind ever. In most cases, however, genetic offspring).Critics might charge that creationists can algorithms are nothing else than probabilistic optimization explain these things without recourse to evolution. For methods which are based on the principles of evolution. Using example, creationists often explain the development of simulation and Genetic Algorithms to improve cluster tool resistance to antibiotic agents in bacteria, or the changes performance, Mooring Pattern Optimization using Genetic wrought in domesticated animals by artificial selection, by Algorithms. This paper is designed to cover a few important application aspects of genetic algorithm under a single umbrella. presuming that God decided to create organisms in fixed Key words: groups, called "kinds" or baramin. Though natural Genetic Algorithm microevolution or human-guided artificial selection can bring about different varieties within the originally created "dog-kind," or "cow-kind," or "bacteria-kind" (!), Introduction no amount of time or genetic change can transform one "kind" into another. However, exactly how the creationists determine what a "kind" is, or what mechanism prevents Creationists occasionally charge that evolution is useless living things from evolving beyond its boundaries, is as a scientific theory because it produces no practical invariably never explained.
    [Show full text]
  • Biogeography-Based Optimization with Blended Migration for Constrained Optimization Problems
    Cleveland State University EngagedScholarship@CSU Electrical Engineering & Computer Science Electrical Engineering & Computer Science Faculty Publications Department 7-2010 Biogeography-Based Optimization with Blended Migration for Constrained Optimization Problems Haiping Ma Shaoxing University, Shaoxing, China, [email protected] Daniel J. Simon Cleveland State University, [email protected] Follow this and additional works at: https://engagedscholarship.csuohio.edu/enece_facpub Part of the Electrical and Computer Engineering Commons How does access to this work benefit ou?y Let us know! Publisher's Statement ACM New York, NY, USA ©2010 Original Citation H. Ma and D. Simon. (2010). Biogeography-Based Optimization with Blended Migration for Constrained Optimization Problems. Genetic and Evolutionary Computation Conference, 417-418. Repository Citation Ma, Haiping and Simon, Daniel J., "Biogeography-Based Optimization with Blended Migration for Constrained Optimization Problems" (2010). Electrical Engineering & Computer Science Faculty Publications. 184. https://engagedscholarship.csuohio.edu/enece_facpub/184 This Conference Proceeding is brought to you for free and open access by the Electrical Engineering & Computer Science Department at EngagedScholarship@CSU. It has been accepted for inclusion in Electrical Engineering & Computer Science Faculty Publications by an authorized administrator of EngagedScholarship@CSU. For more information, please contact [email protected]. Biogeography-Based Optimization with Blended Migration for Constrained Optimization Problems Haiping Ma Dan Simon Shaoxing University Cleveland State University Shaoxing, Zhejiang, 312000, China Cleveland, Ohio [email protected] [email protected] ABSTRACT and which is motivated by blended crossover in GAs [6]. Blended Biogeography-based optimization (BBO) is a new evolutionary migration is defined as algorithm based on the science of biogeography.
    [Show full text]
  • A Short Introduction to Evolutionary Computation (In 50 Minutes)
    A Short Introduction to Evolutionary Computation (in 50 minutes) Fernando Lobo University of Algarve Outline • Why EC in 50 minutes? • The basic idea of a Genetic Algorithm • Application example Why EC in 50 minutes? • If I only had 50 minutes to talk to someone about Evolutionary Computation, this is what I would show them. • Two purposes: – to show you the basic idea quickly, without getting involved in too much details and theory. – to get you very motivated to study this topic. Many variations on EC • There’s a variety of classes of Evolutionary Algorithms – Genetic Algorithms (GAs) – Evolution Strategies (ES) – Evolutionary Programming (EP) – Genetic Programming (GP) – … • In this lecture, we will only look at the so-called Simple Genetic Algorithm. What is a Genetic Algorithm (GA)? • It is a search procedure based on the mechanics of natural selection and genetics. • Often used to solve difficult optimization problems. Evolution • Evolution can be seen as an optimization process where living creatures constantly adapt to their environment. – the stronger survive and propagate their traits to future generations – the weaker die and their traits tend to disappear. • Darwin’s survival of the fittest. How do GAs differ from other techniques? • Work with a population of solutions, rather than a single solution. • Use probabilistic rather than deterministic mechanisms. Two essential components – Selection – Variation • recombination (or crossover) • mutation When should we use a GA? • When other simpler methods are not good enough. • When we don’t have much knowledge about the problem that we are trying to solve. • … and when the search space is very large (making complete enumeration unfeasible).
    [Show full text]
  • A Model of Protocell Based on the Introduction of a Semi-Permeable Membrane in a Stochastic Model of Catalytic Reaction Networks
    A model of protocell based on the introduction of a semi-permeable membrane in a stochastic model of catalytic reaction networks Roberto Serra Marco Villani Department of Physics, Informatics and Mathematics Modena and Reggio Emilia University ECLT European Centre for Living Technology S. Marco 2940, University of Ca’ Foscari, Venice (IT) [email protected] [email protected] Alessandro Filisetti ECLT European Centre for Living Technology S. Marco 2940, University of Ca’ Foscari, Venice (IT) [email protected] Alex Graudenzi Chiara Damiani Department of Informatics Systems and Communication, University of Milan Bicocca [email protected] [email protected] Abstract The theoretical characterization of the self-organizing molecular structures emerging from ensembles of distinct interacting chemicals turns to be very important in revealing those dynamics that led to the transition from the non-living to the living matter as well as in the design of artificial protocells [12, 13, 14]. In this work we aim at studying the role of a semi-permeable membrane, i.e. a very simple protocell description, in the dynamics of a stochastic model describing randomly generated catalytic reaction sets (CRSs) of molecules. Catalytic reaction sets are networks composed of different molecules interacting together leading to the formation of several reaction pathways. In particular, molecules are representation of molecular species formed by a concatenation of letters (bricks) from a finite alphabet. Reaction pathways are formed by two kinds of reaction only, namely condensation, where two molecular species are glued together in order to create a longer species and cleavage where a molecular species is divided in two shorter species, Figure 1, Filisetti et al.
    [Show full text]
  • Genetic Algorithm Based on Natural Selection Theory for Optimization Problems
    S S symmetry Article Genetic Algorithm Based on Natural Selection Theory for Optimization Problems Musatafa Abbas Albadr 1,*, Sabrina Tiun 1, Masri Ayob 1 and Fahad AL-Dhief 2 1 CAIT, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia; [email protected] (S.T.); [email protected] (M.A.) 2 School of Electrical Engineering, Department of Communication Engineering, Universiti Teknologi Malaysia, UTM Johor Bahru, Johor 81310, Malaysia; [email protected] * Correspondence: [email protected]; Tel.: +60-104213847 Received: 29 August 2020; Accepted: 5 October 2020; Published: 23 October 2020 Abstract: The metaheuristic genetic algorithm (GA) is based on the natural selection process that falls under the umbrella category of evolutionary algorithms (EA). Genetic algorithms are typically utilized for generating high-quality solutions for search and optimization problems by depending on bio-oriented operators such as selection, crossover, and mutation. However, the GA still suffers from some downsides and needs to be improved so as to attain greater control of exploitation and exploration concerning creating a new population and randomness involvement happening in the population at the solution initialization. Furthermore, the mutation is imposed upon the new chromosomes and hence prevents the achievement of an optimal solution. Therefore, this study presents a new GA that is centered on the natural selection theory and it aims to improve the control of exploitation and exploration. The proposed algorithm is called genetic algorithm based on natural selection theory (GABONST). Two assessments of the GABONST are carried out via (i) application of fifteen renowned benchmark test functions and the comparison of the results with the conventional GA, enhanced ameliorated teaching learning-based optimization (EATLBO), Bat and Bee algorithms.
    [Show full text]