Theory of Evolutionary Computation

Total Page:16

File Type:pdf, Size:1020Kb

Theory of Evolutionary Computation Tutorial on Evolutionary Techniques and Fuzzy Logic in Power Systems Prepared for the IEEE-PES Summer Meeting in Seattle July, 2000 Course Coordinator: Mohamed A. El-Sharkawi, University of Washington Contributors: Chapter 1: Theory of Evolutionary Computation Russel Eberhart, Purdue School of Engineering and Technology, Indiana University Purdue University Indianapolis Chapter 2: An Introduction to Fuzzy Inference Robert J. Marks, University of Washington Chapter 3: Application of Evolutionary Technique to Power System Security Assessment Craig Jensen, Microsoft Corporation M. A. El-Sharkawi, University of Washington Robert J. Marks, University of Washington Chapter 4: Application of EC to Optimization, Model Identification, Training and Control Alexandre P. Alves da Silva, Federal Engineering School at Itajuba, Brazil Chapter 5: Fuzzy Systems Applications to Power Systems Kevin Tomsovic, Washington State University Evolutionary Techniques and Fuzzy Logic in Power Systems Table of Contents Chapter 1: Theory of Evolutionary Computation .......................................................................................... 4 1.1 Introduction.......................................................................................................................................... 4 1.2 EC Paradigm Attributes ....................................................................................................................... 4 1.3 Implementation .................................................................................................................................... 4 1.4 Genetic Algorithms.............................................................................................................................. 5 1.4.1 Introduction................................................................................................................................... 5 1.4.2 An Overview of Genetic Algorithms ............................................................................................ 5 1.4.3 A Simple GA Example Problem................................................................................................... 6 1.4.4 A Review of GA Operations......................................................................................................... 8 1.5 Particle Swarm Optimization............................................................................................................. 11 1.5.1 Introduction................................................................................................................................. 11 1.5.2 Simulating Social Behavior ........................................................................................................ 12 1.5.3 The Particle Swarm Optimization Concept................................................................................. 13 1.5.4 Training a Multilayer Perceptron................................................................................................ 13 1.6 Conclusions........................................................................................................................................ 15 1.7 Acknowledgments.............................................................................................................................. 16 1.8 Selected Bibliography........................................................................................................................ 16 Chapter 2: An Introduction to Fuzzy Inference ............................................................................................ 19 2.1 Fuzzy Sets.......................................................................................................................................... 19 2.2 Differences Between Fuzzy Membership Functions and Probability Density Functions .................. 19 2.3 Fuzzy Logic ....................................................................................................................................... 21 2.3.1 Fuzzy If-Then Rules ................................................................................................................... 22 2.3.2 Numerical Interpretation of the Fuzzy Antecedent ..................................................................... 22 2.3.3 Matrix Descriptions of Fuzzy Rules ........................................................................................... 24 2.4 Application to Control ....................................................................................................................... 24 2.5 Variations........................................................................................................................................... 25 2.5.1 Alternate Fuzzy Logic................................................................................................................. 25 2.5.2 Defuzzification............................................................................................................................ 25 2.5.3 Weighting Consequents .............................................................................................................. 26 2.6 Further Reading ................................................................................................................................. 26 Chapter 3: Application of Evolutionary Technique to Power System Security Assessment ........................ 27 3.1 Abstract.............................................................................................................................................. 27 3.2 Introduction........................................................................................................................................ 27 3.3 NN's for DSA..................................................................................................................................... 28 3.4 Evolutionary-Based Query Learning Algorithm................................................................................28 3.5 Case Study – IEEE 17 Generator System .......................................................................................... 29 3.6 Conclusions........................................................................................................................................ 30 3.7 References.......................................................................................................................................... 30 Chapter 4: Application of EC to Optimization, Model Identification, Training and Control ....................... 32 4.1 Optimization ...................................................................................................................................... 32 4.1.1 Modern Heuristic Search Techniques .......................................................................................... 32 4.1.2 Power System Applications ........................................................................................................ 33 4.1.3 Example ...................................................................................................................................... 34 4.2 MODEL IDENTIFICATION ............................................................................................................ 34 4.2.1 Dynamic Load Modeling ............................................................................................................ 35 4.2.2 Short-Term Load Forecasting ..................................................................................................... 35 4.3 TRAINING ........................................................................................................................................ 36 4.3.1 Pruning Versus Growing............................................................................................................. 36 4.3.2 Types of Approximation Functions ............................................................................................ 37 4.3.3 The Polynomial Network............................................................................................................ 37 4.4 CONTROL......................................................................................................................................... 38 4.5 ACKNOWLEDGMENTS ................................................................................................................. 38 4.6 REFERENCES .................................................................................................................................. 38 Chapter 5: Fuzzy Systems Applications to Power Systems..........................................................................41 5.1 Introduction........................................................................................................................................ 41 5.2 Power System Applications ............................................................................................................... 42 5.2.1 Rule-based Fuzzy Systems.......................................................................................................... 42 5.2.2 Fuzzy Controllers........................................................................................................................ 43 5.2.3 Fuzzy Decision-Making and Optimization ................................................................................. 43 5.3 Basics of Fuzzy Mathematics ...........................................................................................................
Recommended publications
  • Metaheuristics1
    METAHEURISTICS1 Kenneth Sörensen University of Antwerp, Belgium Fred Glover University of Colorado and OptTek Systems, Inc., USA 1 Definition A metaheuristic is a high-level problem-independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms (Sörensen and Glover, To appear). Notable examples of metaheuristics include genetic/evolutionary algorithms, tabu search, simulated annealing, and ant colony optimization, although many more exist. A problem-specific implementation of a heuristic optimization algorithm according to the guidelines expressed in a metaheuristic framework is also referred to as a metaheuristic. The term was coined by Glover (1986) and combines the Greek prefix meta- (metá, beyond in the sense of high-level) with heuristic (from the Greek heuriskein or euriskein, to search). Metaheuristic algorithms, i.e., optimization methods designed according to the strategies laid out in a metaheuristic framework, are — as the name suggests — always heuristic in nature. This fact distinguishes them from exact methods, that do come with a proof that the optimal solution will be found in a finite (although often prohibitively large) amount of time. Metaheuristics are therefore developed specifically to find a solution that is “good enough” in a computing time that is “small enough”. As a result, they are not subject to combinatorial explosion – the phenomenon where the computing time required to find the optimal solution of NP- hard problems increases as an exponential function of the problem size. Metaheuristics have been demonstrated by the scientific community to be a viable, and often superior, alternative to more traditional (exact) methods of mixed- integer optimization such as branch and bound and dynamic programming.
    [Show full text]
  • Completely Derandomized Self-Adaptation in Evolution Strategies
    Completely Derandomized Self-Adaptation in Evolution Strategies Nikolaus Hansen [email protected] Technische Universitat¨ Berlin, Fachgebiet fur¨ Bionik, Sekr. ACK 1, Ackerstr. 71–76, 13355 Berlin, Germany Andreas Ostermeier [email protected] Technische Universitat¨ Berlin, Fachgebiet fur¨ Bionik, Sekr. ACK 1, Ackerstr. 71–76, 13355 Berlin, Germany Abstract This paper puts forward two useful methods for self-adaptation of the mutation dis- tribution – the concepts of derandomization and cumulation. Principle shortcomings of the concept of mutative strategy parameter control and two levels of derandomization are reviewed. Basic demands on the self-adaptation of arbitrary (normal) mutation dis- tributions are developed. Applying arbitrary, normal mutation distributions is equiv- alent to applying a general, linear problem encoding. The underlying objective of mutative strategy parameter control is roughly to favor previously selected mutation steps in the future. If this objective is pursued rigor- ously, a completely derandomized self-adaptation scheme results, which adapts arbi- trary normal mutation distributions. This scheme, called covariance matrix adaptation (CMA), meets the previously stated demands. It can still be considerably improved by cumulation – utilizing an evolution path rather than single search steps. Simulations on various test functions reveal local and global search properties of the evolution strategy with and without covariance matrix adaptation. Their per- formances are comparable only on perfectly scaled functions. On badly scaled, non- separable functions usually a speed up factor of several orders of magnitude is ob- served. On moderately mis-scaled functions a speed up factor of three to ten can be expected. Keywords Evolution strategy, self-adaptation, strategy parameter control, step size control, de- randomization, derandomized self-adaptation, covariance matrix adaptation, evolu- tion path, cumulation, cumulative path length control.
    [Show full text]
  • A Discipline of Evolutionary Programming 1
    A Discipline of Evolutionary Programming 1 Paul Vit¶anyi 2 CWI, Kruislaan 413, 1098 SJ Amsterdam, The Netherlands. Email: [email protected]; WWW: http://www.cwi.nl/ paulv/ » Genetic ¯tness optimization using small populations or small pop- ulation updates across generations generally su®ers from randomly diverging evolutions. We propose a notion of highly probable ¯t- ness optimization through feasible evolutionary computing runs on small size populations. Based on rapidly mixing Markov chains, the approach pertains to most types of evolutionary genetic algorithms, genetic programming and the like. We establish that for systems having associated rapidly mixing Markov chains and appropriate stationary distributions the new method ¯nds optimal programs (individuals) with probability almost 1. To make the method useful would require a structured design methodology where the develop- ment of the program and the guarantee of the rapidly mixing prop- erty go hand in hand. We analyze a simple example to show that the method is implementable. More signi¯cant examples require theoretical advances, for example with respect to the Metropolis ¯lter. Note: This preliminary version may deviate from the corrected ¯nal published version Theoret. Comp. Sci., 241:1-2 (2000), 3{23. 1 Introduction Performance analysis of genetic computing using unbounded or exponential population sizes or population updates across generations [27,21,25,28,29,22,8] may not be directly applicable to real practical problems where we always have to deal with a bounded (small) population size [9,23,26]. 1 Preliminary version published in: Proc. 7th Int'nl Workshop on Algorithmic Learning Theory, Lecture Notes in Arti¯cial Intelligence, Vol.
    [Show full text]
  • A New Biogeography-Based Optimization Approach Based on Shannon-Wiener Diversity Index to Pid Tuning in Multivariable System
    ABCM Symposium Series in Mechatronics - Vol. 5 Section III – Emerging Technologies and AI Applications Copyright © 2012 by ABCM Page 592 A NEW BIOGEOGRAPHY-BASED OPTIMIZATION APPROACH BASED ON SHANNON-WIENER DIVERSITY INDEX TO PID TUNING IN MULTIVARIABLE SYSTEM Marsil de Athayde Costa e Silva, [email protected] Undergraduate course of Mechatronics Engineering Camila da Costa Silveira, [email protected] Leandro dos Santos Coelho, [email protected] Industrial and Systems Engineering Graduate Program, PPGEPS Pontifical Catholic University of Parana, PUCPR Imaculada Conceição, 1155, Zip code 80215-901, Curitiba, Parana, Brazil Abstract. Proportional-integral-derivative (PID) control is the most popular control architecture used in industrial problems. Many techniques have been proposed to tune the gains for the PID controller. Over the last few years, as an alternative to the conventional mathematical approaches, modern metaheuristics, such as evolutionary computation and swarm intelligence paradigms, have been given much attention by many researchers due to their ability to find good solutions in PID tuning. As a modern metaheuristic method, Biogeography-based optimization (BBO) is a generalization of biogeography to evolutionary algorithm inspired on the mathematical model of organism distribution in biological systems. BBO is an evolutionary process that achieves information sharing by biogeography-based migration operators. This paper proposes a modification for the BBO using a diversity index, called Shannon-wiener index (SW-BBO), for tune the gains of the PID controller in a multivariable system. Results show that the proposed SW-BBO approach is efficient to obtain high quality solutions in PID tuning. Keywords: PID control, biogeography-based optimization, Shannon-Wiener index, multivariable systems.
    [Show full text]
  • Evolutionary Algorithms
    Evolutionary Algorithms Dr. Sascha Lange AG Maschinelles Lernen und Naturlichsprachliche¨ Systeme Albert-Ludwigs-Universit¨at Freiburg [email protected] Dr. Sascha Lange Machine Learning Lab, University of Freiburg Evolutionary Algorithms (1) Acknowlegements and Further Reading These slides are mainly based on the following three sources: I A. E. Eiben, J. E. Smith, Introduction to Evolutionary Computing, corrected reprint, Springer, 2007 — recommendable, easy to read but somewhat lengthy I B. Hammer, Softcomputing,LectureNotes,UniversityofOsnabruck,¨ 2003 — shorter, more research oriented overview I T. Mitchell, Machine Learning, McGraw Hill, 1997 — very condensed introduction with only a few selected topics Further sources include several research papers (a few important and / or interesting are explicitly cited in the slides) and own experiences with the methods described in these slides. Dr. Sascha Lange Machine Learning Lab, University of Freiburg Evolutionary Algorithms (2) ‘Evolutionary Algorithms’ (EA) constitute a collection of methods that originally have been developed to solve combinatorial optimization problems. They adapt Darwinian principles to automated problem solving. Nowadays, Evolutionary Algorithms is a subset of Evolutionary Computation that itself is a subfield of Artificial Intelligence / Computational Intelligence. Evolutionary Algorithms are those metaheuristic optimization algorithms from Evolutionary Computation that are population-based and are inspired by natural evolution.Typicalingredientsare: I A population (set) of individuals (the candidate solutions) I Aproblem-specificfitness (objective function to be optimized) I Mechanisms for selection, recombination and mutation (search strategy) There is an ongoing controversy whether or not EA can be considered a machine learning technique. They have been deemed as ‘uninformed search’ and failing in the sense of learning from experience (‘never make an error twice’).
    [Show full text]
  • Genetic Algorithms for Applied Path Planning a Thesis Presented in Partial Ful Llment of the Honors Program Vincent R
    Genetic Algorithms for Applied Path Planning A Thesis Presented in Partial Ful llment of the Honors Program Vincent R. Ragusa Abstract Path planning is the computational task of choosing a path through an environment. As a task humans do hundreds of times a day, it may seem that path planning is an easy task, and perhaps naturally suited for a computer to solve. This is not the case however. There are many ways in which NP-Hard problems like path planning can be made easier for computers to solve, but the most signi cant of these is the use of approximation algorithms. One such approximation algorithm is called a genetic algorithm. Genetic algorithms belong to a an area of computer science called evolutionary computation. The techniques used in evolutionary computation algorithms are modeled after the principles of Darwinian evolution by natural selection. Solutions to the problem are literally bred for their problem solving ability through many generations of selective breeding. The goal of the research presented is to examine the viability of genetic algorithms as a practical solution to the path planning problem. Various modi cations to a well known genetic algorithm (NSGA-II) were implemented and tested experimentally to determine if the modi cation had an e ect on the operational eciency of the algorithm. Two new forms of crossover were implemented with positive results. The notion of mass extinction driving evolution was tested with inconclusive results. A path correction algorithm called make valid was created which has proven to be extremely powerful. Finally several additional objective functions were tested including a path smoothness measure and an obstacle intrusion measure, the latter showing an enormous positive result.
    [Show full text]
  • Evolutionary Programming Made Faster
    82 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 3, NO. 2, JULY 1999 Evolutionary Programming Made Faster Xin Yao, Senior Member, IEEE, Yong Liu, Student Member, IEEE, and Guangming Lin Abstract— Evolutionary programming (EP) has been applied used to generate new solutions (offspring) and selection is used with success to many numerical and combinatorial optimization to test which of the newly generated solutions should survive problems in recent years. EP has rather slow convergence rates, to the next generation. Formulating EP as a special case of the however, on some function optimization problems. In this paper, a “fast EP” (FEP) is proposed which uses a Cauchy instead generate-and-test method establishes a bridge between EP and of Gaussian mutation as the primary search operator. The re- other search algorithms, such as evolution strategies, genetic lationship between FEP and classical EP (CEP) is similar to algorithms, simulated annealing (SA), tabu search (TS), and that between fast simulated annealing and the classical version. others, and thus facilitates cross-fertilization among different Both analytical and empirical studies have been carried out to research areas. evaluate the performance of FEP and CEP for different function optimization problems. This paper shows that FEP is very good One disadvantage of EP in solving some of the multimodal at search in a large neighborhood while CEP is better at search optimization problems is its slow convergence to a good in a small local neighborhood. For a suite of 23 benchmark near-optimum (e.g., to studied in this paper). The problems, FEP performs much better than CEP for multimodal generate-and-test formulation of EP indicates that mutation functions with many local minima while being comparable to is a key search operator which generates new solutions from CEP in performance for unimodal and multimodal functions with only a few local minima.
    [Show full text]
  • The Roles of Evolutionary Computation, Fitness Landscape, Constructive Methods and Local Searches in the Development of Adaptive Systems for Infrastructure Planning
    International Symposium for Next Generation Infrastructure October 1-4, 2013, Wollongong, Australia The roles of evolutionary computation, fitness landscape, constructive methods and local searches in the development of adaptive systems for infrastructure planning Mehrdad Amirghasemi a* Reza Zamani a Abstract: Modelling and systems simulation for improving city planning, and traffic equilibrium are involved with the development of adaptive systems that can learn and respond to the environment intelligently. By employing sophisticated techniques which highly support infrastructure planning and design, evolutionary computation can play a key role in the development of such systems. The key to presenting solution strategies for these systems is fitness landscape which makes some problems hard and some problems easy to tackle. Moreover, constructive methods and local searches can assist evolutionary searches to improve their performance. In this paper, in the context of infrastructure, in general, and city planning, and traffic equilibrium, in particular, the integration of these four concepts is briefly discussed. Key words: Evolutionary computation; Local search; Infrastructure; City planning; Traffic equilibrium. I. Introduction Infrastructure is a term used to indicate both organizational and physical structures a society or an enterprise needs to function. In effect, all facilities and services needed for the operation of an economy is gathered under the umbrella term of “Infrastructure”. Each Infrastructure includes connected elements which are structurally related and each element can affect the other elements. It is through this interconnection that these elements collectively provide a basis for structural development of a society or an enterprise. Viewed in an operative level, infrastructure makes the production of goods and services required in the society more straightforward and the distribution of finished products to market easier.
    [Show full text]
  • Neuroevolution Strategies for Episodic Reinforcement Learning ∗ Verena Heidrich-Meisner , Christian Igel
    J. Algorithms 64 (2009) 152–168 Contents lists available at ScienceDirect Journal of Algorithms Cognition, Informatics and Logic www.elsevier.com/locate/jalgor Neuroevolution strategies for episodic reinforcement learning ∗ Verena Heidrich-Meisner , Christian Igel Institut für Neuroinformatik, Ruhr-Universität Bochum, 44780 Bochum, Germany article info abstract Article history: Because of their convincing performance, there is a growing interest in using evolutionary Received 30 April 2009 algorithms for reinforcement learning. We propose learning of neural network policies Available online 8 May 2009 by the covariance matrix adaptation evolution strategy (CMA-ES), a randomized variable- metric search algorithm for continuous optimization. We argue that this approach, Keywords: which we refer to as CMA Neuroevolution Strategy (CMA-NeuroES), is ideally suited for Reinforcement learning Evolution strategy reinforcement learning, in particular because it is based on ranking policies (and therefore Covariance matrix adaptation robust against noise), efficiently detects correlations between parameters, and infers a Partially observable Markov decision process search direction from scalar reinforcement signals. We evaluate the CMA-NeuroES on Direct policy search five different (Markovian and non-Markovian) variants of the common pole balancing problem. The results are compared to those described in a recent study covering several RL algorithms, and the CMA-NeuroES shows the overall best performance. © 2009 Elsevier Inc. All rights reserved. 1. Introduction Neuroevolution denotes the application of evolutionary algorithms to the design and learning of neural networks [54,11]. Accordingly, the term Neuroevolution Strategies (NeuroESs) refers to evolution strategies applied to neural networks. Evolution strategies are a major branch of evolutionary algorithms [35,40,8,2,7].
    [Show full text]
  • Sustainable Growth and Synchronization in Protocell Models
    life Article Sustainable Growth and Synchronization in Protocell Models Roberto Serra 1,2,3 and Marco Villani 1,2,* 1 Department of Physics, Informatics and Mathematics, Modena and Reggio Emilia University, Via Campi 213/A, 41125 Modena, Italy 2 European Centre for Living Technology, Ca’ Bottacin, Dorsoduro 3911, Calle Crosera, 30123 Venice, Italy 3 Institute for Advanced Study, University of Amsterdam, Oude Turfmarkt 147, 1012 GC Amsterdam, The Netherlands * Correspondence: [email protected] Received: 19 July 2019; Accepted: 14 August 2019; Published: 21 August 2019 Abstract: The growth of a population of protocells requires that the two key processes of replication of the protogenetic material and reproduction of the whole protocell take place at the same rate. While in many ODE-based models such synchronization spontaneously develops, this does not happen in the important case of quadratic growth terms. Here we show that spontaneous synchronization can be recovered (i) by requiring that the transmembrane diffusion of precursors takes place at a finite rate, or (ii) by introducing a finite lifetime of the molecular complexes. We then consider reaction networks that grow by the addition of newly synthesized chemicals in a binary polymer model, and analyze their behaviors in growing and dividing protocells, thereby confirming the importance of (i) and (ii) for synchronization. We describe some interesting phenomena (like long-term oscillations of duplication times) and show that the presence of food-generated autocatalytic cycles is not sufficient to guarantee synchronization: in the case of cycles with a complex structure, it is often observed that only some subcycles survive and synchronize, while others die out.
    [Show full text]
  • Arxiv:1803.03453V4 [Cs.NE] 21 Nov 2019
    The Surprising Creativity of Digital Evolution: A Collection of Anecdotes from the Evolutionary Computation and Artificial Life Research Communities Joel Lehman1†, Jeff Clune1, 2†, Dusan Misevic3†, Christoph Adami4, Lee Altenberg5, Julie Beaulieu6, Peter J Bentley7, Samuel Bernard8, Guillaume Beslon9, David M Bryson4, Patryk Chrabaszcz11, Nick Cheney2, Antoine Cully12, Stephane Doncieux13, Fred C Dyer4, Kai Olav Ellefsen14, Robert Feldt15, Stephan Fischer16, Stephanie Forrest17, Antoine Fr´enoy18, Christian Gagn´e6 Leni Le Goff13, Laura M Grabowski19, Babak Hodjat20, Frank Hutter11, Laurent Keller21, Carole Knibbe9, Peter Krcah22, Richard E Lenski4, Hod Lipson23, Robert MacCurdy24, Carlos Maestre13, Risto Miikkulainen26, Sara Mitri21, David E Moriarty27, Jean-Baptiste Mouret28, Anh Nguyen2, Charles Ofria4, Marc Parizeau 6, David Parsons9, Robert T Pennock4, William F Punch4, Thomas S Ray29, Marc Schoenauer30, Eric Schulte17, Karl Sims, Kenneth O Stanley1,31, Fran¸coisTaddei3, Danesh Tarapore32, Simon Thibault6, Westley Weimer33, Richard Watson34, Jason Yosinski1 †Organizing lead authors 1 Uber AI Labs, San Francisco, CA, USA 2 University of Wyoming, Laramie, WY, USA 3 Center for Research and Interdisciplinarity, Paris, France 4 Michigan State University, East Lansing, MI, USA 5 Univeristy of Hawai‘i at Manoa, HI, USA 6 Universit´eLaval, Quebec City, Quebec, Canada 7 University College London, London, UK 8 INRIA, Institut Camille Jordan, CNRS, UMR5208, 69622 Villeurbanne, France 9 Universit´ede Lyon, INRIA, CNRS, LIRIS UMR5205, INSA, UCBL,
    [Show full text]
  • A Note on Evolutionary Algorithms and Its Applications
    Bhargava, S. (2013). A Note on Evolutionary Algorithms and Its Applications. Adults Learning Mathematics: An International Journal, 8(1), 31-45 A Note on Evolutionary Algorithms and Its Applications Shifali Bhargava Dept. of Mathematics, B.S.A. College, Mathura (U.P)- India. <[email protected]> Abstract This paper introduces evolutionary algorithms with its applications in multi-objective optimization. Here elitist and non-elitist multiobjective evolutionary algorithms are discussed with their advantages and disadvantages. We also discuss constrained multiobjective evolutionary algorithms and their applications in various areas. Key words: evolutionary algorithms, multi-objective optimization, pareto-optimality, elitist. Introduction The term evolutionary algorithm (EA) stands for a class of stochastic optimization methods that simulate the process of natural evolution. The origins of EAs can be traced back to the late 1950s, and since the 1970’s several evolutionary methodologies have been proposed, mainly genetic algorithms, evolutionary programming, and evolution strategies. All of these approaches operate on a set of candidate solutions. Using strong simplifications, this set is subsequently modified by the two basic principles of evolution: selection and variation. Selection represents the competition for resources among living beings. Some are better than others and more likely to survive and to reproduce their genetic information. In evolutionary algorithms, natural selection is simulated by a stochastic selection process. Each solution is given a chance to reproduce a certain number of times, dependent on their quality. Thereby, quality is assessed by evaluating the individuals and assigning them scalar fitness values. The other principle, variation, imitates natural capability of creating “new” living beings by means of recombination and mutation.
    [Show full text]