A Natural Evolution Strategy with Asynchronous Strategy Updates

Total Page:16

File Type:pdf, Size:1020Kb

A Natural Evolution Strategy with Asynchronous Strategy Updates A Natural Evolution Strategy with Asynchronous Strategy Updates Tobias Glasmachers Institut für Neuroinformatik Ruhr-Universität Bochum, Germany [email protected] ABSTRACT since it is a standard assumption that the lion’s share of the We propose a generic method for turning a modern, non- overall computation time is spent on this step. The pos- elitist evolution strategy with fully adaptive covariance ma- sibility to perform individual steps asynchronously enables trix into an asynchronous algorithm. This algorithm can more efficient parallelism, since synchronization points form process the result of an evaluation of the fitness function potential bottlenecks. anytime and update its search strategy, without the need Evolution Strategies (ESs) are special within the wider to synchronize with the rest of the population. The asyn- field of evolutionary computation in that they rely heavily chronous update builds on the recent developments of natu- on active“self”-adaptation of their search or mutation distri- ral evolution strategies and information geometric optimiza- bution. While this technique is a blessing in many respects tion. (it enables linear convergence into twice continuously dif- Our algorithm improves on the usual generational scheme ferentiable optima), it is also a curse, since it can interfere in two respects. Remarkably, the possibility to process fit- with typical difficulties of evolutionary optimization, such ness values immediately results in a speed-up of the sequen- as fitness noise and constraint handling. The same holds tial algorithm. Furthermore, our algorithm is much better for asynchronous processing. In the present study we show suited for parallel processing. It allows to use more proces- how to make fitness evaluation in a modern non-elitist evo- sors than offspring individuals in a meaningful way. lution strategy an asynchronous operation, while preserving meaningful strategy updates of the full covariance structure. Parallelization of evolutionary algorithms (EAs) has been Categories and Subject Descriptors subject to a large number of studies since EAs, being popula- [Evolution Strategies and Evolutionary Programming] tion-based algorithms, are well suited for parallelization. Different types of parallelism have been identified, see e.g. [1]. General Terms Relatively simple master-slave architectures can distribute fitness evaluations within an offspring population to multiple Algorithms processors. More complex systems are based on distributed or otherwise structured populations, like for example island Keywords models [2] with synchronous or asynchronous message pass- Evolution strategies, Speedup technique, Parallelization ing. Recently there have been impressive demonstrations of 1. INTRODUCTION massively parallel evolutionary algorithms using huge pop- ulation sizes of tens of thousands [7]. Such implementation In pure form, most evolutionary algorithms (EAs) oper- rely heavily of the general purpose computing capabilities ate in generation cycles. For many specific variants it is of modern graphics processing units (GPUs). Such mas- easy to weaken this assumption and to allow for anytime sively parallel implementations are most commonly found or asynchronous application of certain operators, such as for genetic algorithms (GAs) and genetic programming (GP) selection of parents, generation of offspring (application of systems, where the search distribution is defined by static variation operators), as well as the evaluation of offspring, operators (without self-adaptation) applied to the current an operation that is typically considered expensive. Other population. operators basically require synchronous operation, such as In general, parallelism and synchronity of different steps survivor selection. Among these, the evaluation of individ- on an algorithm are distinct properties. However, asyn- uals by means of the fitness function is of primary interest, chronicity of an operation is useless in a strictly sequential program, while it can avoid synchronization overheads in a parallel computation. Asynchronicity is not a prerequisite Permission to make digital or hard copies of all or part of this work for for parallelism, but it can increase its efficiency. personal or classroom use is granted without fee provided that copies are In this paper we do not aim for the massive parallelism not made or distributed for profit or commercial advantage and that copies mentioned above. Speed-ups in the order of hundreds are bear this notice and the full citation on the first page. To copy otherwise, to not to be expected for (todays) ESs, simply for their com- republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. paratively small population sizes. Therefore, instead of GPU GECCO’13, July 6-10, 2013, Amsterdam, The Netherlands. hardware, we consider a setting that is better fitted for the Copyright 2013 ACM 978-1-4503-1963-8/13/07 ...$15.00. standard ES: assume the fitness evaluation involves a com- tially become stochastic gradient steps in parameter space, putation intensive simulation that needs a full blown CPU- which are augmented by learning rates. based architecture to run efficiently. We have a few dedi- Our approach modifies a generational ES so that it be- cated compute servers available for the task, with a total comes a fully asynchronous algorithm. The asynchronous of c independent processors, with c in the order of maybe scheme creates an offspring xi as soon as a compute node dozens. Then we can choose any multiple of c for the off- becomes idle. This compute node is then busy for some time spring population size n to obtain a basically linear speed- evaluating the offspring’s fitness f(xi). In general, fitness up, compared to the naive sequential implementation. This values may be returned in an arbitrary order by the com- can be achieved with a simple master-slave architecture. It pute nodes, so that evaluated offspring (xi,f(xi)) become is assumed in particular that the lion’s share of the com- available in an order that is different from the generation of putation time is spent on fitness evaluations, not on com- offspring. In this setting we argue that evaluated individuals munication overhead, strategy updates, sampling of random should not be viewed as members of populations in a discrete numbers, and relatively cheap bookkeeping tasks that are cycle, but rather as a continuous stream of individuals. performed after synchronization by the central master pro- This view offers the opportunity to use all information as cess. Such work has been conducted, e.g., for the highly soon as they become available for updating the search distri- efficient CMA-ES algorithm [5]. bution from which new offspring are generated. Now assume Although we have already made quite a few assumptions that a compute node has just returned the fitness value of at this point, our performance calculation can still turn out an individual. Put in a nutshell, our approach amounts to to be very far off. For example, it may well happen that applying a 1/n fraction of the strategy update of a stan- the ES would actually run at its highest efficiency with a dard generational ES, to account for the higher update fre- smaller population size. While this effect is often small (al- quency, using the n most recently arrived individuals from though considerable for huge populations, see e.g. [5]), an- the stream as the current population. This rule needs minor other one may turn out to be crucial: the runtime of the corrections to account for noise induced by delayed arrival simulation may vary unpredictably (with the search point, of individuals due to varying evaluations times. with the compute node, with factors deeply hidden in indus- Our simple yet efficient scheme can deal with variation in trial simulation software, or with other factors outside our the runtimes of fitness evaluations without wasting compu- control). Then most processors will have to wait for the syn- tation time. Interestingly it also improves sequential search chronization with the slowest-to-evaluate individual in the on a single core. The reason is that the first individual of a population. hypothetical offspring population is readily evaluated before A principled solution to this problem is to break up the the second one needs to be generated. The information con- generation cycle and to turn to asynchronous algorithms. tained in the first fitness value can already guide the search The synchronous generation cycle of many EAs is obviously to better points, even within a single generation. an over-simplification of natural evolution, and there has The remainder of this paper is organized as follows: Nat- been considerable work to overcome this restriction algo- ural evolution strategies in general and the xNES algorithm rithmically [2]. in particular are presented in the context of information ge- The standard solution to the asynchronous update re- ometric optimization. Then we introduce the asynchronous quirement in EAs is steady state selection. The elitist char- update. The new algorithm is benchmarked in a number of acter of selection schemes directly suitable for asynchronicity different sequential and parallel settings against the popula- is problematic in the presence of multi-modality and noise tion-based ES. We close with our conclusions. in fitness values, which are both common characteristics of fitness functions based on simulations. 2. NATURAL EVOLUTION STRATEGIES Furthermore, most existing schemes are designed for
Recommended publications
  • Metaheuristics1
    METAHEURISTICS1 Kenneth Sörensen University of Antwerp, Belgium Fred Glover University of Colorado and OptTek Systems, Inc., USA 1 Definition A metaheuristic is a high-level problem-independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms (Sörensen and Glover, To appear). Notable examples of metaheuristics include genetic/evolutionary algorithms, tabu search, simulated annealing, and ant colony optimization, although many more exist. A problem-specific implementation of a heuristic optimization algorithm according to the guidelines expressed in a metaheuristic framework is also referred to as a metaheuristic. The term was coined by Glover (1986) and combines the Greek prefix meta- (metá, beyond in the sense of high-level) with heuristic (from the Greek heuriskein or euriskein, to search). Metaheuristic algorithms, i.e., optimization methods designed according to the strategies laid out in a metaheuristic framework, are — as the name suggests — always heuristic in nature. This fact distinguishes them from exact methods, that do come with a proof that the optimal solution will be found in a finite (although often prohibitively large) amount of time. Metaheuristics are therefore developed specifically to find a solution that is “good enough” in a computing time that is “small enough”. As a result, they are not subject to combinatorial explosion – the phenomenon where the computing time required to find the optimal solution of NP- hard problems increases as an exponential function of the problem size. Metaheuristics have been demonstrated by the scientific community to be a viable, and often superior, alternative to more traditional (exact) methods of mixed- integer optimization such as branch and bound and dynamic programming.
    [Show full text]
  • Covariance Matrix Adaptation for the Rapid Illumination of Behavior Space
    Covariance Matrix Adaptation for the Rapid Illumination of Behavior Space Matthew C. Fontaine Julian Togelius Viterbi School of Engineering Tandon School of Engineering University of Southern California New York University Los Angeles, CA New York City, NY [email protected] [email protected] Stefanos Nikolaidis Amy K. Hoover Viterbi School of Engineering Ying Wu College of Computing University of Southern California New Jersey Institute of Technology Los Angeles, CA Newark, NJ [email protected] [email protected] ABSTRACT We focus on the challenge of finding a diverse collection of quality solutions on complex continuous domains. While quality diver- sity (QD) algorithms like Novelty Search with Local Competition (NSLC) and MAP-Elites are designed to generate a diverse range of solutions, these algorithms require a large number of evaluations for exploration of continuous spaces. Meanwhile, variants of the Covariance Matrix Adaptation Evolution Strategy (CMA-ES) are among the best-performing derivative-free optimizers in single- objective continuous domains. This paper proposes a new QD algo- rithm called Covariance Matrix Adaptation MAP-Elites (CMA-ME). Figure 1: Comparing Hearthstone Archives. Sample archives Our new algorithm combines the self-adaptation techniques of for both MAP-Elites and CMA-ME from the Hearthstone ex- CMA-ES with archiving and mapping techniques for maintaining periment. Our new method, CMA-ME, both fills more cells diversity in QD. Results from experiments based on standard con- in behavior space and finds higher quality policies to play tinuous optimization benchmarks show that CMA-ME finds better- Hearthstone than MAP-Elites. Each grid cell is an elite (high quality solutions than MAP-Elites; similarly, results on the strategic performing policy) and the intensity value represent the game Hearthstone show that CMA-ME finds both a higher overall win rate across 200 games against difficult opponents.
    [Show full text]
  • Improving Fitness Functions in Genetic Programming for Classification on Unbalanced Credit Card Datasets
    Improving Fitness Functions in Genetic Programming for Classification on Unbalanced Credit Card Datasets Van Loi Cao, Nhien An Le Khac, Miguel Nicolau, Michael O’Neill, James McDermott University College Dublin Abstract. Credit card fraud detection based on machine learning has recently attracted considerable interest from the research community. One of the most important tasks in this area is the ability of classifiers to handle the imbalance in credit card data. In this scenario, classifiers tend to yield poor accuracy on the fraud class (minority class) despite realizing high overall accuracy. This is due to the influence of the major- ity class on traditional training criteria. In this paper, we aim to apply genetic programming to address this issue by adapting existing fitness functions. We examine two fitness functions from previous studies and develop two new fitness functions to evolve GP classifier with superior accuracy on the minority class and overall. Two UCI credit card datasets are used to evaluate the effectiveness of the proposed fitness functions. The results demonstrate that the proposed fitness functions augment GP classifiers, encouraging fitter solutions on both the minority and the majority classes. 1 Introduction Credit cards have emerged as the preferred means of payment in response to continuous economic growth and rapid developments in information technology. The daily volume of credit card transactions reveals the shift from cash to card. Concurrently, the incidence of credit card fraud has risen with the loss of billions of dollars globally each year. Moreover, criminals have evolved increasingly more sophisticated strategies to exploit weaknesses in the security protocols.
    [Show full text]
  • A Discipline of Evolutionary Programming 1
    A Discipline of Evolutionary Programming 1 Paul Vit¶anyi 2 CWI, Kruislaan 413, 1098 SJ Amsterdam, The Netherlands. Email: [email protected]; WWW: http://www.cwi.nl/ paulv/ » Genetic ¯tness optimization using small populations or small pop- ulation updates across generations generally su®ers from randomly diverging evolutions. We propose a notion of highly probable ¯t- ness optimization through feasible evolutionary computing runs on small size populations. Based on rapidly mixing Markov chains, the approach pertains to most types of evolutionary genetic algorithms, genetic programming and the like. We establish that for systems having associated rapidly mixing Markov chains and appropriate stationary distributions the new method ¯nds optimal programs (individuals) with probability almost 1. To make the method useful would require a structured design methodology where the develop- ment of the program and the guarantee of the rapidly mixing prop- erty go hand in hand. We analyze a simple example to show that the method is implementable. More signi¯cant examples require theoretical advances, for example with respect to the Metropolis ¯lter. Note: This preliminary version may deviate from the corrected ¯nal published version Theoret. Comp. Sci., 241:1-2 (2000), 3{23. 1 Introduction Performance analysis of genetic computing using unbounded or exponential population sizes or population updates across generations [27,21,25,28,29,22,8] may not be directly applicable to real practical problems where we always have to deal with a bounded (small) population size [9,23,26]. 1 Preliminary version published in: Proc. 7th Int'nl Workshop on Algorithmic Learning Theory, Lecture Notes in Arti¯cial Intelligence, Vol.
    [Show full text]
  • Evolutionary Algorithms
    Evolutionary Algorithms Dr. Sascha Lange AG Maschinelles Lernen und Naturlichsprachliche¨ Systeme Albert-Ludwigs-Universit¨at Freiburg [email protected] Dr. Sascha Lange Machine Learning Lab, University of Freiburg Evolutionary Algorithms (1) Acknowlegements and Further Reading These slides are mainly based on the following three sources: I A. E. Eiben, J. E. Smith, Introduction to Evolutionary Computing, corrected reprint, Springer, 2007 — recommendable, easy to read but somewhat lengthy I B. Hammer, Softcomputing,LectureNotes,UniversityofOsnabruck,¨ 2003 — shorter, more research oriented overview I T. Mitchell, Machine Learning, McGraw Hill, 1997 — very condensed introduction with only a few selected topics Further sources include several research papers (a few important and / or interesting are explicitly cited in the slides) and own experiences with the methods described in these slides. Dr. Sascha Lange Machine Learning Lab, University of Freiburg Evolutionary Algorithms (2) ‘Evolutionary Algorithms’ (EA) constitute a collection of methods that originally have been developed to solve combinatorial optimization problems. They adapt Darwinian principles to automated problem solving. Nowadays, Evolutionary Algorithms is a subset of Evolutionary Computation that itself is a subfield of Artificial Intelligence / Computational Intelligence. Evolutionary Algorithms are those metaheuristic optimization algorithms from Evolutionary Computation that are population-based and are inspired by natural evolution.Typicalingredientsare: I A population (set) of individuals (the candidate solutions) I Aproblem-specificfitness (objective function to be optimized) I Mechanisms for selection, recombination and mutation (search strategy) There is an ongoing controversy whether or not EA can be considered a machine learning technique. They have been deemed as ‘uninformed search’ and failing in the sense of learning from experience (‘never make an error twice’).
    [Show full text]
  • Accelerating Evolutionary Algorithms with Gaussian Process Fitness Function Models Dirk Buche,¨ Nicol N
    IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS — PART C: APPLICATIONS AND REVIEWS, VOL. 35, NO. 2, MAY 2005 183 Accelerating Evolutionary Algorithms with Gaussian Process Fitness Function Models Dirk Buche,¨ Nicol N. Schraudolph, and Petros Koumoutsakos Abstract— We present an overview of evolutionary algorithms II. MODELS IN EVOLUTIONARY ALGORITHMS that use empirical models of the fitness function to accelerate convergence, distinguishing between evolution control and the There are two main ways to integrate models into an surrogate approach. We describe the Gaussian process model evolutionary optimization. In the first, a fraction of individuals and propose using it as an inexpensive fitness function surrogate. is evaluated on the fitness function itself, the remainder Implementation issues such as efficient and numerically stable merely on its model. Jin et al. [7] refer to individuals that computation, exploration versus exploitation, local modeling, are evaluated on the fitness function as controlled, and call multiple objectives and constraints, and failed evaluations are ad- dressed. Our resulting Gaussian process optimization procedure this technique evolution control. In the second approach, the clearly outperforms other evolutionary strategies on standard test optimum of the model is determined, then evaluated on the functions as well as on a real-world problem: the optimization fitness function. The new evaluation is used to update the of stationary gas turbine compressor profiles. model, and the process is repeated with the improved model. Index Terms— evolutionary algorithms, fitness function mod- This surrogate approach evaluates only predicted optima on eling, evolution control, surrogate approach, Gaussian process, the fitness function, otherwise using the model as a surrogate.
    [Show full text]
  • Evolutionary Programming Made Faster
    82 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 3, NO. 2, JULY 1999 Evolutionary Programming Made Faster Xin Yao, Senior Member, IEEE, Yong Liu, Student Member, IEEE, and Guangming Lin Abstract— Evolutionary programming (EP) has been applied used to generate new solutions (offspring) and selection is used with success to many numerical and combinatorial optimization to test which of the newly generated solutions should survive problems in recent years. EP has rather slow convergence rates, to the next generation. Formulating EP as a special case of the however, on some function optimization problems. In this paper, a “fast EP” (FEP) is proposed which uses a Cauchy instead generate-and-test method establishes a bridge between EP and of Gaussian mutation as the primary search operator. The re- other search algorithms, such as evolution strategies, genetic lationship between FEP and classical EP (CEP) is similar to algorithms, simulated annealing (SA), tabu search (TS), and that between fast simulated annealing and the classical version. others, and thus facilitates cross-fertilization among different Both analytical and empirical studies have been carried out to research areas. evaluate the performance of FEP and CEP for different function optimization problems. This paper shows that FEP is very good One disadvantage of EP in solving some of the multimodal at search in a large neighborhood while CEP is better at search optimization problems is its slow convergence to a good in a small local neighborhood. For a suite of 23 benchmark near-optimum (e.g., to studied in this paper). The problems, FEP performs much better than CEP for multimodal generate-and-test formulation of EP indicates that mutation functions with many local minima while being comparable to is a key search operator which generates new solutions from CEP in performance for unimodal and multimodal functions with only a few local minima.
    [Show full text]
  • Neuroevolution Strategies for Episodic Reinforcement Learning ∗ Verena Heidrich-Meisner , Christian Igel
    J. Algorithms 64 (2009) 152–168 Contents lists available at ScienceDirect Journal of Algorithms Cognition, Informatics and Logic www.elsevier.com/locate/jalgor Neuroevolution strategies for episodic reinforcement learning ∗ Verena Heidrich-Meisner , Christian Igel Institut für Neuroinformatik, Ruhr-Universität Bochum, 44780 Bochum, Germany article info abstract Article history: Because of their convincing performance, there is a growing interest in using evolutionary Received 30 April 2009 algorithms for reinforcement learning. We propose learning of neural network policies Available online 8 May 2009 by the covariance matrix adaptation evolution strategy (CMA-ES), a randomized variable- metric search algorithm for continuous optimization. We argue that this approach, Keywords: which we refer to as CMA Neuroevolution Strategy (CMA-NeuroES), is ideally suited for Reinforcement learning Evolution strategy reinforcement learning, in particular because it is based on ranking policies (and therefore Covariance matrix adaptation robust against noise), efficiently detects correlations between parameters, and infers a Partially observable Markov decision process search direction from scalar reinforcement signals. We evaluate the CMA-NeuroES on Direct policy search five different (Markovian and non-Markovian) variants of the common pole balancing problem. The results are compared to those described in a recent study covering several RL algorithms, and the CMA-NeuroES shows the overall best performance. © 2009 Elsevier Inc. All rights reserved. 1. Introduction Neuroevolution denotes the application of evolutionary algorithms to the design and learning of neural networks [54,11]. Accordingly, the term Neuroevolution Strategies (NeuroESs) refers to evolution strategies applied to neural networks. Evolution strategies are a major branch of evolutionary algorithms [35,40,8,2,7].
    [Show full text]
  • A Note on Evolutionary Algorithms and Its Applications
    Bhargava, S. (2013). A Note on Evolutionary Algorithms and Its Applications. Adults Learning Mathematics: An International Journal, 8(1), 31-45 A Note on Evolutionary Algorithms and Its Applications Shifali Bhargava Dept. of Mathematics, B.S.A. College, Mathura (U.P)- India. <[email protected]> Abstract This paper introduces evolutionary algorithms with its applications in multi-objective optimization. Here elitist and non-elitist multiobjective evolutionary algorithms are discussed with their advantages and disadvantages. We also discuss constrained multiobjective evolutionary algorithms and their applications in various areas. Key words: evolutionary algorithms, multi-objective optimization, pareto-optimality, elitist. Introduction The term evolutionary algorithm (EA) stands for a class of stochastic optimization methods that simulate the process of natural evolution. The origins of EAs can be traced back to the late 1950s, and since the 1970’s several evolutionary methodologies have been proposed, mainly genetic algorithms, evolutionary programming, and evolution strategies. All of these approaches operate on a set of candidate solutions. Using strong simplifications, this set is subsequently modified by the two basic principles of evolution: selection and variation. Selection represents the competition for resources among living beings. Some are better than others and more likely to survive and to reproduce their genetic information. In evolutionary algorithms, natural selection is simulated by a stochastic selection process. Each solution is given a chance to reproduce a certain number of times, dependent on their quality. Thereby, quality is assessed by evaluating the individuals and assigning them scalar fitness values. The other principle, variation, imitates natural capability of creating “new” living beings by means of recombination and mutation.
    [Show full text]
  • Analysis of Evolutionary Algorithms
    Why We Do Not Evolve Software? Analysis of Evolutionary Algorithms Roman V. Yampolskiy Computer Engineering and Computer Science University of Louisville [email protected] Abstract In this paper, we review the state-of-the-art results in evolutionary computation and observe that we don’t evolve non-trivial software from scratch and with no human intervention. A number of possible explanations are considered, but we conclude that computational complexity of the problem prevents it from being solved as currently attempted. A detailed analysis of necessary and available computational resources is provided to support our findings. Keywords: Darwinian Algorithm, Genetic Algorithm, Genetic Programming, Optimization. “We point out a curious philosophical implication of the algorithmic perspective: if the origin of life is identified with the transition from trivial to non-trivial information processing – e.g. from something akin to a Turing machine capable of a single (or limited set of) computation(s) to a universal Turing machine capable of constructing any computable object (within a universality class) – then a precise point of transition from non-life to life may actually be undecidable in the logical sense. This would likely have very important philosophical implications, particularly in our interpretation of life as a predictable outcome of physical law.” [1]. 1. Introduction On April 1, 2016 Dr. Yampolskiy posted the following to his social media accounts: “Google just announced major layoffs of programmers. Future software development and updates will be done mostly via recursive self-improvement by evolving deep neural networks”. The joke got a number of “likes” but also, interestingly, a few requests from journalists for interviews on this “developing story”.
    [Show full text]
  • Evolutionary Algorithms and Their Applications to Engineering Problems
    Neural Computing and Applications https://doi.org/10.1007/s00521-020-04832-8 (0123456789().,-volV)(0123456789().,- volV) REVIEW ARTICLE Evolutionary algorithms and their applications to engineering problems 1 2 Adam Slowik • Halina Kwasnicka Received: 27 November 2018 / Accepted: 5 March 2020 Ó The Author(s) 2020 Abstract The main focus of this paper is on the family of evolutionary algorithms and their real-life applications. We present the following algorithms: genetic algorithms, genetic programming, differential evolution, evolution strategies, and evolu- tionary programming. Each technique is presented in the pseudo-code form, which can be used for its easy implementation in any programming language. We present the main properties of each algorithm described in this paper. We also show many state-of-the-art practical applications and modifications of the early evolutionary methods. The open research issues are indicated for the family of evolutionary algorithms. Keywords Nature-inspired methods Á Genetic algorithm Á Genetic programming Á Differential evolution Á Evolution strategy Á Evolutionary programming Á Real-life applications 1 Introduction This paper is a state-of-the-art paper which topic is connected mainly with evolutionary algorithms (EAs) such These days, in the area of soft computing research we can as GA, GP, DE, ES, and EP. (In the other paper [6], we observe a strong pressure to search for new optimization have presented swarm intelligence algorithms (SIAs) such techniques which are based on nature. Figure 1 presents as ant colony optimization (ACO), particle swarm opti- some approaches in optimization techniques with a con- mization (PSO), and others in which social collaboration centration on evolutionary approaches.
    [Show full text]
  • Evolutionary Theories of Aging"
    Article ID: WMC003275 ISSN 2046-1690 The Germ-soma Conflict Theory of Aging and Death: Obituary to the "Evolutionary Theories of Aging" Corresponding Author: Prof. Kurt Heininger, Professor, Department of Neurology, Heinrich Heine University Duesseldorf - Germany Submitting Author: Prof. Kurt Heininger, Professor, Department of Neurology, Heinrich Heine University Duesseldorf - Germany Article ID: WMC003275 Article Type: Original Articles Submitted on:20-Apr-2012, 01:51:59 PM GMT Published on: 20-Apr-2012, 05:38:59 PM GMT Article URL: http://www.webmedcentral.com/article_view/3275 Subject Categories:AGING Keywords:Aging, Reproduction, Evolution, Darwinian demon, Resources How to cite the article:Heininger K. The Germ-soma Conflict Theory of Aging and Death: Obituary to the "Evolutionary Theories of Aging" . WebmedCentral AGING 2012;3(4):WMC003275 Copyright: This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Source(s) of Funding: No funding Competing Interests: No competing interests Additional Files: References WebmedCentral > Original Articles Page 1 of 139 WMC003275 Downloaded from http://www.webmedcentral.com on 30-Apr-2012, 10:35:59 AM The Germ-soma Conflict Theory of Aging and Death: Obituary to the "Evolutionary Theories of Aging" Author(s): Heininger K Abstract “cheater” phenotypes that are defective to carry the cost of death at reproductive events are less fit under conditions of feast and famine cycles and have a high risk to go extinct. Tracing both the genomic and Regarding aging, the scientific community persists in a physiological “fossil record” and the phenotypic pattern state of collective schizophrenia.
    [Show full text]