List of Symbols

Total Page:16

File Type:pdf, Size:1020Kb

List of Symbols List of Symbols V solution space for the global optimization problem d(·, ·) distance function in the space V dH (·, ·) Hamming distance function · 2 Euclidean norm in the space V · arbitrary norm in the space V N dimension of the global optimization problem (N =dim(V )) D admissible set of solutions to the global optimization problem Φ : D→R+ objective function of a maximization problem Φ˜ : D→R+ objective function of a minimization problem x∗ global maximizer (or minimizer) to the global optimization problem x+ local maximizer (or minimizer) to the global optimization problem meas(·) Lesbegue measure function Loc set of local optimization methods loc(·) local optimization method Rloc attractor of the local minimizer x+ with respect to the strictly x+ decreasing local optimization method loc(·) B + x+ basin of attraction of the local minimizer x P random sample, population Pr(·) probability measure M(A) space of probabilistic measures over the set A b(·) function that selects the best fitted individual in a population Dr ⊂D grid of encoded points (set of phenotypes) r =#Dr the cardinality of phenotype set U genetic universum (set of genotypes) code : U →Dr encoding bijective function dcode : D −→ U partial function of inverse encoding (decoding partial function) Ω binary genetic universum in Chapters 3 - 5, space of elementary events in Section 6.1.5 l binary code length Z2 group {0, 1} with the addition modulo 2 204 List of Symbols ⊕ the addition operator in Z2, coordinate-by-coordinate addition of binary vectors from Z2×,...,×Z2 codea(·) affine binary encoding codeG(·) Gray binary encoding f : U → R+ fitness function Scale(·) nonlinear function that modifies fitness µ number of individuals in the parental population self (·) probability distribution of selection Elite subset of individuals that pass to the next epoch with the probability 1 1 vector of l ones ˆi the inverse of the binary vector i (ˆi = 1 ⊕ i) pm rate of binary mutation pc rate of binary crossover [·] evaluation function for boolean expressions ⊗ coordinate-by-coordinate multiplication of binary vectors type type of binary crossover mutx(·) probability distribution for the binary mutation of an individual x crossx,y(·) probability distribution for the binary crossover of individuals x and y N (e, C) realization (result of sampling) of the N-dimensional Gauss random variable with the mean vector e and the covariance matrix C N (e, σ) realization (result of sampling) of the one dimensional Gauss random variable with the mean value e and the standard deviation σ U[0, 1] realization (result of sampling) of the random variable of the uniform probability distribution over the real interval [0, 1] λ number of offspring individuals in a single genetic epoch κ individual life time parameter of the evolutionary algorithm in section 5.3.2, raster resolution in section 6.1.10 S space of schemata H schemata from the space S ∆(H) length of the schemata H ℵ(H) degree of the schemata H E(·) expected value operator E space of states of genetic algorithms eqp equivalence relation among vectors of genotypes Sµ group of permutations of the µ-element set Λr−1 unit r − 1 simplex in Rr r−1 Xµ ⊂ Λ finite set of states of the genetic algorithm with finite population (µ<+∞) #A cardinality of the set or multiset A n =#E cardinality of the space of states πt genetic algorithm state probability distribution in the epoch t τ Markov transition function of genetic algorithm states Q probability transition matrix F (·) selection operator for the simple genetic algorithm List of Symbols 205 M(·) mixing operator for the simple genetic algorithm G(·) genetic (heuristics) operator for the simple genetic algorithm M mixing matrix K⊂Λr−1 set of fixed points of the genetic operator G dim(V ) dimension of the vector space V diam(A) diameter of the subset A of a metric space N set of natural numbers Z set of integers Z+ set of nonnegative integers Z+ = N ∪{0} R set of rational numbers R+ set of nonnegative rational numbers top(V ) topology on V (the family of open sets in V ) A the topological closure of the set A in the proper topology (A)− the complement of the set A, i.e. (A)− = V \ A if it is contained in the space V χA the characteristic function of the set A. Assuming A iscontainedinthespaceV χA : V →{0, 1},χA(x)=1if x ∈ A, χA(x)=0otherwise · operator of the upper round “ceiling” · operator of the lower round “floor” Int : R → N function turning back the nearest integer to the argument int(A) interior of the set A in the proper topology 2 diag : RN → RN function turning back the square diagonal matrix diag(v) with the diagonal equal to the vector v supp(g) support of the real valued function g : A → R, supp(g)={x ∈ A; g(x) =0 } I identity mapping I the matrix of linear identity of finite dimensional vector space Dom(f) domain of the function f u(t) vector of parameters that control genetic operations in the epoch t L(Rr → Rr) space of linear mappings from Rr into itself Γ Gamma Euler’s function References 1. Ali M, Storey C (1994) Topographical Multilevel Single Linkage. Journal of Global Optimization 5:349–358. 2. Anderson RW (1997) The Baldwin Effect. Chapter C.3.4.1 in [15]. 3. Anderssen RS, Bloomfield P (1975) Properties of the random search in global optimization. Journal of Optimization Theory and Applications 16:383–389. 4. Arabas J (1995) Evolutionary Algorithms with the varying population cardi- nality and the varying crossover range. PhD Thesis, Warsaw University of Technology, Warsaw (in Polish). 5. Arabas J (2001) Lectures in Evolutionary Algorithms. WNT, Warsaw (in Pol- ish). 6. Arabas J (2003) Sampling measure of an evolutionary algorithm. In: Proc. of 6-th KAEiOG Conf., Łagów Lubuski, 15–20. 7. Arabas J, Michalewicz Z, Mulawka J (1994) GAVaPS – a Genetic Algorithm with Varying Population Size. Proc.ofICEC’94, Orlando, Florida, IEEE Press, 73–76. 8. Arabas J, Słomka M (2000) Pseudo-random number generators in the initial population generation. In: Proc. of 4-th KAEiOG Conf., Lądek Zdrój, 7–12 (in Polish). 9. Archetti Betrò B (1978) On the Effectiveness of Uniform Random Sampling in Global Optimization Problems. Technical Report, University of Pisa. 10. Bäck T, Fogel DB, Michalewicz Z eds. (2000) Evolutionary Computation 1. Basic Algorithms and Operators. Institute of Physics Publishing, Bristol and Philadelphia. 11. Bäck T, Fogel DB, Michalewicz Z eds. (2000) Evolutionary Computation 2. Advanced Algorithms and Operators. Institute of Physics Publishing, Bristol and Philadelphia. 12. Bäck T (1996) Evolutionary Algorithms in Theory and Practice. Oxford Univ. Press. 13. Bäck T (1997) Self–adaptation. Chapter C.7.1 in [15]. 14. Bäck T (1997) Mutation Parameters. Chapter E.1.2 in [15]. 15. Bäck T, Fogel DB, Michalewicz Z (1997) Handbook of Evolutionary Compu- tations. Oxford University Press. 16. Bäck T, Schütz M (1996) Intelligent Mutation Note Control in Canonical Ge- netic Algorithm. Foundation of Intelligent Systems. In: Ras ZW, Michalewicz Zeds.Proc. of 9-th Int. Symp. ISIM’96., LNCS 1079, Springer. 208 References 17. Bagley JD (1967) The Behavior of Adaptive Systems with Employ Genetic and Correlation Algorithms. PhD Thesis, University of Michigan, Dissertation Abstracts International 28(12), 5106B. (Univ. Microfilms No. 68–7556). 18. Bahadur RR (1966) A Note on Quantiles in Large Samples. Annals of Math- ematical Statistics 37:577–580. 19. Beasley D, Bull DR, Martin RR (1993) A Sequential Niche for Multimodal Function Optimization. Evolutionary Computation Vol. 1, No. 2, 101–125. 20. Becker RW, Lago GV (1970) A Global Optimization Algorithm. In: Proc. of Allerton Conf. on Circuits and System Theory, Monticallo, Illinois, 3–15. 21. Bethke AD (1981) Genetic Algorithms as Function Optimizers. PhD Thesis, University of Michigan, Dissertation Abstracts International 41(9), 3503B, (Univ. Microfilms No. 8106101). 22. Betrò B (1981) Bayesian Testing of Nonparametric Hypotheses and its Appli- cation to Global Optimization. Technical Reports CNR-IAMI, Italy. 23. Betrò, B Schoen F(1987) Sequential Stopping Rules for the Multistart Algo- rithm in Global Optimization. Mathematical Programming 38:271–286. 24. Betrò, B Schoen F(1992) Sequential Stopping Rules for the Multistart Algo- rithm in Global Optimization. Mathematical Programming 52:445–458. 25. Beyer HG (1995) Toward a Theory of Evolution Strategies: Self-adoption. Evolutionary Computation 3:311–348. 26. Beyer HG (2001) The Theory of Evolution Strategies. Springer. 27. Beyer HG, Rudolph G (1997) Local Performance Measures. Chapter B.2.4. in [15]. 28. Billingsley P (1979) Probability and Measure. John Willey and Sons, New York, Chichester, Brisbane, Toronto. 29. Boender CGE (1984) The Generalized Multinominal Distribution: A Bayesian Analysis and Applications. PhD Thesis, Erasmus University, Rotterdam, Cen- trum voor Wiskunde en Informatica, Amsterdam. 30. Boender CGE, Rinnooy Kan AHG (1991) On when to Stop Sampling for the Maximum. Journal of Global Optimization 1:331–340. 31. Boender CGE, Rinnoy Kan AHG (1985) Bayesian Stopping Rules for a Class of Stochastic Global Optimization Methods. Technical Report, Econometric Institute, Erasmus University, Rotterdam. 32. Boender CGE, Rinnoy Kan AHG (1987) Bayesian Stopping Rules for Multi- start Global Optimization Methods. Mathematical Programming 37:59–80. 33. Boender CGE, Rinnoy Kan AHG, Vercellis C (1987) Stochastic Optimization. In: Andreatta G, Mason F, Serafini P eds. Advanced School on Stochastics in Combinatorial Optimization. Word Scientific, Singapore. 34. Boender CGE, Rinnoy Kan AHG, Stougie L, Timmer GT (1982) A Stochastic Method for Global Optimization. Mathematical Programming 22:125–140. 35. Boender CGE, Zieliński R (1985) A Sequential Bayesian Approach to Esti- mating the Dimension of a Multinominal Distribution. In: Sequential Methods in Statistics. Banach Center Publications Vol.
Recommended publications
  • Metaheuristics1
    METAHEURISTICS1 Kenneth Sörensen University of Antwerp, Belgium Fred Glover University of Colorado and OptTek Systems, Inc., USA 1 Definition A metaheuristic is a high-level problem-independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms (Sörensen and Glover, To appear). Notable examples of metaheuristics include genetic/evolutionary algorithms, tabu search, simulated annealing, and ant colony optimization, although many more exist. A problem-specific implementation of a heuristic optimization algorithm according to the guidelines expressed in a metaheuristic framework is also referred to as a metaheuristic. The term was coined by Glover (1986) and combines the Greek prefix meta- (metá, beyond in the sense of high-level) with heuristic (from the Greek heuriskein or euriskein, to search). Metaheuristic algorithms, i.e., optimization methods designed according to the strategies laid out in a metaheuristic framework, are — as the name suggests — always heuristic in nature. This fact distinguishes them from exact methods, that do come with a proof that the optimal solution will be found in a finite (although often prohibitively large) amount of time. Metaheuristics are therefore developed specifically to find a solution that is “good enough” in a computing time that is “small enough”. As a result, they are not subject to combinatorial explosion – the phenomenon where the computing time required to find the optimal solution of NP- hard problems increases as an exponential function of the problem size. Metaheuristics have been demonstrated by the scientific community to be a viable, and often superior, alternative to more traditional (exact) methods of mixed- integer optimization such as branch and bound and dynamic programming.
    [Show full text]
  • A Discipline of Evolutionary Programming 1
    A Discipline of Evolutionary Programming 1 Paul Vit¶anyi 2 CWI, Kruislaan 413, 1098 SJ Amsterdam, The Netherlands. Email: [email protected]; WWW: http://www.cwi.nl/ paulv/ » Genetic ¯tness optimization using small populations or small pop- ulation updates across generations generally su®ers from randomly diverging evolutions. We propose a notion of highly probable ¯t- ness optimization through feasible evolutionary computing runs on small size populations. Based on rapidly mixing Markov chains, the approach pertains to most types of evolutionary genetic algorithms, genetic programming and the like. We establish that for systems having associated rapidly mixing Markov chains and appropriate stationary distributions the new method ¯nds optimal programs (individuals) with probability almost 1. To make the method useful would require a structured design methodology where the develop- ment of the program and the guarantee of the rapidly mixing prop- erty go hand in hand. We analyze a simple example to show that the method is implementable. More signi¯cant examples require theoretical advances, for example with respect to the Metropolis ¯lter. Note: This preliminary version may deviate from the corrected ¯nal published version Theoret. Comp. Sci., 241:1-2 (2000), 3{23. 1 Introduction Performance analysis of genetic computing using unbounded or exponential population sizes or population updates across generations [27,21,25,28,29,22,8] may not be directly applicable to real practical problems where we always have to deal with a bounded (small) population size [9,23,26]. 1 Preliminary version published in: Proc. 7th Int'nl Workshop on Algorithmic Learning Theory, Lecture Notes in Arti¯cial Intelligence, Vol.
    [Show full text]
  • Evolutionary Algorithms
    Evolutionary Algorithms Dr. Sascha Lange AG Maschinelles Lernen und Naturlichsprachliche¨ Systeme Albert-Ludwigs-Universit¨at Freiburg [email protected] Dr. Sascha Lange Machine Learning Lab, University of Freiburg Evolutionary Algorithms (1) Acknowlegements and Further Reading These slides are mainly based on the following three sources: I A. E. Eiben, J. E. Smith, Introduction to Evolutionary Computing, corrected reprint, Springer, 2007 — recommendable, easy to read but somewhat lengthy I B. Hammer, Softcomputing,LectureNotes,UniversityofOsnabruck,¨ 2003 — shorter, more research oriented overview I T. Mitchell, Machine Learning, McGraw Hill, 1997 — very condensed introduction with only a few selected topics Further sources include several research papers (a few important and / or interesting are explicitly cited in the slides) and own experiences with the methods described in these slides. Dr. Sascha Lange Machine Learning Lab, University of Freiburg Evolutionary Algorithms (2) ‘Evolutionary Algorithms’ (EA) constitute a collection of methods that originally have been developed to solve combinatorial optimization problems. They adapt Darwinian principles to automated problem solving. Nowadays, Evolutionary Algorithms is a subset of Evolutionary Computation that itself is a subfield of Artificial Intelligence / Computational Intelligence. Evolutionary Algorithms are those metaheuristic optimization algorithms from Evolutionary Computation that are population-based and are inspired by natural evolution.Typicalingredientsare: I A population (set) of individuals (the candidate solutions) I Aproblem-specificfitness (objective function to be optimized) I Mechanisms for selection, recombination and mutation (search strategy) There is an ongoing controversy whether or not EA can be considered a machine learning technique. They have been deemed as ‘uninformed search’ and failing in the sense of learning from experience (‘never make an error twice’).
    [Show full text]
  • Evolutionary Programming Made Faster
    82 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 3, NO. 2, JULY 1999 Evolutionary Programming Made Faster Xin Yao, Senior Member, IEEE, Yong Liu, Student Member, IEEE, and Guangming Lin Abstract— Evolutionary programming (EP) has been applied used to generate new solutions (offspring) and selection is used with success to many numerical and combinatorial optimization to test which of the newly generated solutions should survive problems in recent years. EP has rather slow convergence rates, to the next generation. Formulating EP as a special case of the however, on some function optimization problems. In this paper, a “fast EP” (FEP) is proposed which uses a Cauchy instead generate-and-test method establishes a bridge between EP and of Gaussian mutation as the primary search operator. The re- other search algorithms, such as evolution strategies, genetic lationship between FEP and classical EP (CEP) is similar to algorithms, simulated annealing (SA), tabu search (TS), and that between fast simulated annealing and the classical version. others, and thus facilitates cross-fertilization among different Both analytical and empirical studies have been carried out to research areas. evaluate the performance of FEP and CEP for different function optimization problems. This paper shows that FEP is very good One disadvantage of EP in solving some of the multimodal at search in a large neighborhood while CEP is better at search optimization problems is its slow convergence to a good in a small local neighborhood. For a suite of 23 benchmark near-optimum (e.g., to studied in this paper). The problems, FEP performs much better than CEP for multimodal generate-and-test formulation of EP indicates that mutation functions with many local minima while being comparable to is a key search operator which generates new solutions from CEP in performance for unimodal and multimodal functions with only a few local minima.
    [Show full text]
  • Neuroevolution Strategies for Episodic Reinforcement Learning ∗ Verena Heidrich-Meisner , Christian Igel
    J. Algorithms 64 (2009) 152–168 Contents lists available at ScienceDirect Journal of Algorithms Cognition, Informatics and Logic www.elsevier.com/locate/jalgor Neuroevolution strategies for episodic reinforcement learning ∗ Verena Heidrich-Meisner , Christian Igel Institut für Neuroinformatik, Ruhr-Universität Bochum, 44780 Bochum, Germany article info abstract Article history: Because of their convincing performance, there is a growing interest in using evolutionary Received 30 April 2009 algorithms for reinforcement learning. We propose learning of neural network policies Available online 8 May 2009 by the covariance matrix adaptation evolution strategy (CMA-ES), a randomized variable- metric search algorithm for continuous optimization. We argue that this approach, Keywords: which we refer to as CMA Neuroevolution Strategy (CMA-NeuroES), is ideally suited for Reinforcement learning Evolution strategy reinforcement learning, in particular because it is based on ranking policies (and therefore Covariance matrix adaptation robust against noise), efficiently detects correlations between parameters, and infers a Partially observable Markov decision process search direction from scalar reinforcement signals. We evaluate the CMA-NeuroES on Direct policy search five different (Markovian and non-Markovian) variants of the common pole balancing problem. The results are compared to those described in a recent study covering several RL algorithms, and the CMA-NeuroES shows the overall best performance. © 2009 Elsevier Inc. All rights reserved. 1. Introduction Neuroevolution denotes the application of evolutionary algorithms to the design and learning of neural networks [54,11]. Accordingly, the term Neuroevolution Strategies (NeuroESs) refers to evolution strategies applied to neural networks. Evolution strategies are a major branch of evolutionary algorithms [35,40,8,2,7].
    [Show full text]
  • A Note on Evolutionary Algorithms and Its Applications
    Bhargava, S. (2013). A Note on Evolutionary Algorithms and Its Applications. Adults Learning Mathematics: An International Journal, 8(1), 31-45 A Note on Evolutionary Algorithms and Its Applications Shifali Bhargava Dept. of Mathematics, B.S.A. College, Mathura (U.P)- India. <[email protected]> Abstract This paper introduces evolutionary algorithms with its applications in multi-objective optimization. Here elitist and non-elitist multiobjective evolutionary algorithms are discussed with their advantages and disadvantages. We also discuss constrained multiobjective evolutionary algorithms and their applications in various areas. Key words: evolutionary algorithms, multi-objective optimization, pareto-optimality, elitist. Introduction The term evolutionary algorithm (EA) stands for a class of stochastic optimization methods that simulate the process of natural evolution. The origins of EAs can be traced back to the late 1950s, and since the 1970’s several evolutionary methodologies have been proposed, mainly genetic algorithms, evolutionary programming, and evolution strategies. All of these approaches operate on a set of candidate solutions. Using strong simplifications, this set is subsequently modified by the two basic principles of evolution: selection and variation. Selection represents the competition for resources among living beings. Some are better than others and more likely to survive and to reproduce their genetic information. In evolutionary algorithms, natural selection is simulated by a stochastic selection process. Each solution is given a chance to reproduce a certain number of times, dependent on their quality. Thereby, quality is assessed by evaluating the individuals and assigning them scalar fitness values. The other principle, variation, imitates natural capability of creating “new” living beings by means of recombination and mutation.
    [Show full text]
  • Analysis of Evolutionary Algorithms
    Why We Do Not Evolve Software? Analysis of Evolutionary Algorithms Roman V. Yampolskiy Computer Engineering and Computer Science University of Louisville [email protected] Abstract In this paper, we review the state-of-the-art results in evolutionary computation and observe that we don’t evolve non-trivial software from scratch and with no human intervention. A number of possible explanations are considered, but we conclude that computational complexity of the problem prevents it from being solved as currently attempted. A detailed analysis of necessary and available computational resources is provided to support our findings. Keywords: Darwinian Algorithm, Genetic Algorithm, Genetic Programming, Optimization. “We point out a curious philosophical implication of the algorithmic perspective: if the origin of life is identified with the transition from trivial to non-trivial information processing – e.g. from something akin to a Turing machine capable of a single (or limited set of) computation(s) to a universal Turing machine capable of constructing any computable object (within a universality class) – then a precise point of transition from non-life to life may actually be undecidable in the logical sense. This would likely have very important philosophical implications, particularly in our interpretation of life as a predictable outcome of physical law.” [1]. 1. Introduction On April 1, 2016 Dr. Yampolskiy posted the following to his social media accounts: “Google just announced major layoffs of programmers. Future software development and updates will be done mostly via recursive self-improvement by evolving deep neural networks”. The joke got a number of “likes” but also, interestingly, a few requests from journalists for interviews on this “developing story”.
    [Show full text]
  • Evolutionary Algorithms and Their Applications to Engineering Problems
    Neural Computing and Applications https://doi.org/10.1007/s00521-020-04832-8 (0123456789().,-volV)(0123456789().,- volV) REVIEW ARTICLE Evolutionary algorithms and their applications to engineering problems 1 2 Adam Slowik • Halina Kwasnicka Received: 27 November 2018 / Accepted: 5 March 2020 Ó The Author(s) 2020 Abstract The main focus of this paper is on the family of evolutionary algorithms and their real-life applications. We present the following algorithms: genetic algorithms, genetic programming, differential evolution, evolution strategies, and evolu- tionary programming. Each technique is presented in the pseudo-code form, which can be used for its easy implementation in any programming language. We present the main properties of each algorithm described in this paper. We also show many state-of-the-art practical applications and modifications of the early evolutionary methods. The open research issues are indicated for the family of evolutionary algorithms. Keywords Nature-inspired methods Á Genetic algorithm Á Genetic programming Á Differential evolution Á Evolution strategy Á Evolutionary programming Á Real-life applications 1 Introduction This paper is a state-of-the-art paper which topic is connected mainly with evolutionary algorithms (EAs) such These days, in the area of soft computing research we can as GA, GP, DE, ES, and EP. (In the other paper [6], we observe a strong pressure to search for new optimization have presented swarm intelligence algorithms (SIAs) such techniques which are based on nature. Figure 1 presents as ant colony optimization (ACO), particle swarm opti- some approaches in optimization techniques with a con- mization (PSO), and others in which social collaboration centration on evolutionary approaches.
    [Show full text]
  • Evolutionary Theories of Aging"
    Article ID: WMC003275 ISSN 2046-1690 The Germ-soma Conflict Theory of Aging and Death: Obituary to the "Evolutionary Theories of Aging" Corresponding Author: Prof. Kurt Heininger, Professor, Department of Neurology, Heinrich Heine University Duesseldorf - Germany Submitting Author: Prof. Kurt Heininger, Professor, Department of Neurology, Heinrich Heine University Duesseldorf - Germany Article ID: WMC003275 Article Type: Original Articles Submitted on:20-Apr-2012, 01:51:59 PM GMT Published on: 20-Apr-2012, 05:38:59 PM GMT Article URL: http://www.webmedcentral.com/article_view/3275 Subject Categories:AGING Keywords:Aging, Reproduction, Evolution, Darwinian demon, Resources How to cite the article:Heininger K. The Germ-soma Conflict Theory of Aging and Death: Obituary to the "Evolutionary Theories of Aging" . WebmedCentral AGING 2012;3(4):WMC003275 Copyright: This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Source(s) of Funding: No funding Competing Interests: No competing interests Additional Files: References WebmedCentral > Original Articles Page 1 of 139 WMC003275 Downloaded from http://www.webmedcentral.com on 30-Apr-2012, 10:35:59 AM The Germ-soma Conflict Theory of Aging and Death: Obituary to the "Evolutionary Theories of Aging" Author(s): Heininger K Abstract “cheater” phenotypes that are defective to carry the cost of death at reproductive events are less fit under conditions of feast and famine cycles and have a high risk to go extinct. Tracing both the genomic and Regarding aging, the scientific community persists in a physiological “fossil record” and the phenotypic pattern state of collective schizophrenia.
    [Show full text]
  • On Comparison Between Evolutionary Programming Network Based Learning and Novel Evolution Strategy Algorithm-Based Learning
    Proceedings of the ICEECE December 22-24 (2003), Dhaka, Bangladesh On Comparison between Evolutionary Programming Network based Learning and Novel Evolution Strategy Algorithm-based Learning M. A. Khayer Azad, Md. Shafiqul Islam and M. M. A. Hashem1 Department of Computer Science and Engineering Khulna University of Engineering and Technology Khulna 9203, Bangladesh ABSTRACT This paper presents two different evolutionary systems - Evolutionary Programming Network (EPNet) [1] and Novel Evolutions Strategy (NES) Algorithm [2]. EPNet does both training and architecture evolution simultaneously, whereas NES does a fixed network and only trains the network. Five mutation operators proposed in EPNet to reflect the emphasis on evolving ANN’s behaviors. Close behavioral links between parents and their offspring are maintained by various mutations, such as partial training and node splitting. On the other hand, NES uses two new genetic operators - subpopulation-based max-mean arithmetical crossover and time-variant mutation. The above-mentioned two algorithms have been tested on a number of benchmark problems, such as the medical diagnosis problems (breast cancer, diabetes, and heart disease). The results and the comparison between them are also present in this paper. 1. INTRODUCTION optimal architecture has been found. The other is the simultaneous evolution of both architectures and Now-a-days, neural network is used in. many weights. It is clear that the evolution of pure applications such as robotics, missile projection, and architectures has difficulties in evaluating fitness rocket even in cooker. A number of algorithms are accurately. As a result, the evolution would be developed to train the network and evolve the inefficient. architecture.
    [Show full text]
  • A Brief Survey on Intelligent Swarm-Based Algorithms for Solving Optimization Problems 49
    DOI: 10.5772/intechopen.76979 ProvisionalChapter chapter 4 A Brief Survey onon IntelligentIntelligent Swarm-BasedSwarm-Based AlgorithmsAlgorithms for Solving OptimizationOptimization ProblemsProblems Siew Mooi Lim and Kuan Yew Leong Additional information isis available atat thethe endend ofof thethe chapterchapter http://dx.doi.org/10.5772/intechopen.76979 Abstract This chapter presents an overview of optimization techniques followed by a brief survey on several swarm-based natural inspired algorithms which were introduced in the last decade. These techniques were inspired by the natural processes of plants, foraging behaviors of insects and social behaviors of animals. These swam intelligent methods have been tested on various standard benchmark problems and are capable in solving a wide range of optimization issues including stochastic, robust and dynamic problems. Keywords: optimization, artificial intelligence, swarm intelligence, nature-inspired and bio-inspired computation 1. Introduction Optimization is a form of mathematical procedure for determining optimal allocation of scare resources. In recent years, the optimization area has received enormous attention primarily due to the rapid emerging science and technology in computing, communication, engineering, environment and society. Several types of optimization problems exist. Two important classes of objects for most optimization problems are limited resources and activities. Resources include land size, plant capacity and sales force. Whereas production activities are like produce stainless steel, low carbon steel or high carbon steel; how we solve them will depend on the circumstances to determine the best condition of activity levels using the resources available. All optimization problems have an objective function, constraints, and choice variables which will lead to the improvement in application or audience.
    [Show full text]
  • Covariance Matrix Adapted Evolution Strategy (CMAES)
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Directory of Open Access Journals ISSN: 2229-6956(ONLINE) ICTACT JOURNAL ON SOFT COMPUTING, OCTOBER 2012, VOLUME: 03, ISSUE: 01 APPLICATION OF RESTART COVARIANCE MATRIX ADAPTATION EVOLUTION STRATEGY (RCMA-ES) TO GENERATION EXPANSION PLANNING PROBLEM K. Karthikeyan1, S. Kannan2, S. Baskar3, and C. Thangaraj4 1,2Department of Electrical and Electronics Engineering, Kalasalingam University, India E-mail: [email protected] and [email protected] 3Department of Electrical and Electronics Engineering, Thiagarajar College of Engineering, India E-mail: [email protected] 4Anna University of Technology Chennai, India E-mail: [email protected] Abstract The Covariance Matrix Adaptation Evolution Strategy This paper describes the application of an evolutionary algorithm, (CMA-ES) is a robust stochastic evolutionary algorithm for Restart Covariance Matrix Adaptation Evolution Strategy (RCMA- difficult non-linear non-convex optimization problems [15]. The ES) to the Generation Expansion Planning (GEP) problem. RCMA- CMA-ES is typically applied to unconstrained or bounded ES is a class of continuous Evolutionary Algorithm (EA) derived from constraint optimization problems, and it efficiently minimizes the concept of self-adaptation in evolution strategies, which adapts the unimodal objective functions and is in particularly superior for covariance matrix of a multivariate normal search distribution. The ill-conditioned and non-separable problems [15], [16]. Anne original GEP problem is modified by incorporating Virtual Mapping Procedure (VMP). The GEP problem of a synthetic test systems for 6- Auger and Nikolaus Hansen showed that increasing the year, 14-year and 24-year planning horizons having five types of population size improves the performance of the CMA-ES on candidate units is considered.
    [Show full text]