<<

Mälardalen University Press Licentiate Theses No. 246

Mälardalen University Press Licentiate Theses No. 246 ENHANCING DIFFERENTIAL EVOLUTION FOR SOLVING CONTINUOUS OPTIMIZATION PROBLEMS

ENHANCING DIFFERENTIAL EVOLUTION ALGORITHM FOR SOLVING CONTINUOUS OPTIMIZATION PROBLEMS

Miguel Leon

Miguel2016 Leon

2016

School of Innovation, Design and Engineering

School of Innovation, Design and Engineering Copyright © Miguel Leon, 2016 ISBN 978-91-7485-299-8 ISSN 1651-9256 Printed by Arkitektkopia, Västerås, Sweden Abstract

Differential Evolution (DE) has become one of the most important metaheuris- tics during the recent years, obtaining attractive results in solving many engi- neering optimization problems. However, the performance of DE is not always strong when seeking optimal solutions. It has two major problems in real world applications. First, it can easily get stuck in a local optimum or fail to generate better solutions before the population has converged. Secondly, its performance is significantly influenced by the control parameters, which are problem depen- dent and which vary in different regions of space under exploration. It usually entails a time consuming trial-and-error procedure to set suitable parameters for DE in a specific problem, particularly for those practioners with limited knowledge and experience of using this technique. This thesis aims to develop new DE to address the two afore- mentioned problems. To mitigate the first problem, we studied the hybridiza- tion of DE with local search techniques to enhance the efficiency of search. The main idea is to apply a local search mechanism to the best individual in each generation of DE to exploit the most promising regions during the evo- lutionary processs so as to speed up the convergence or increase the chance to scape from local optima. Four local search strategies have been integrated and tested in the global DE framework, leading to variants of the memetic DE algo- rithms with different properties concerning diversification and intensification. For tackling the second problem, we propose a greedy adaptation method for dynamic adjustment of the control parameters in DE. It is implemented by con- ducting greedy search repeatedly during the run of DE to reach better parameter assignments in the neighborhood of a current candidate. The candidates are as- sessed by considering both, the success rate and also fitness improvement of trial solutions against the target ones. The incorporation of this greedy parame- ter adaptation method into standard DE has led to a new adaptive DE algorithm, referred to as Greedy Adaptive Differential Evolution (GADE). The methods proposed in this thesis have been tested in different bench- mark problems and compared with the state of the art algorithms, obtaining

i ii competitive results. Furthermore, the proposed GADE algorithm has been ap- plied in an industrial scenario achieving more accurate results than those ob- tained by a standard DE algorithm. Sammanfattning

Differential Evolution (DE) har blivit en av de viktigaste metaheuristikerna un- der de senaste aren˚ och har uppntt attraktiva resultat for¨ att losa¨ manga˚ optimer- ingsproblem inom teknik. Dock ar¨ prestationen hos DE inte alltid framgangsrik˚ nar¨ man soker¨ optimala losningar.¨ Det finns tva˚ huvudsakliga problem for¨ ap- plikationer i den verkliga varlden.¨ Det forsta¨ ar¨ att den latt¨ kan fastna i lokala optimum eller misslyckas att generera battre¨ losningar¨ fore¨ det att populatio- nen (en grupp av losningar)¨ har hunnit konvergera. Det andra ar¨ att prestan- dan paverkas˚ markv¨ ardigt¨ av kontrollparametrar, vilkas optimala varden¨ beror pa˚ problem som ska losas¨ och varierar mellan regioner i sokrymden.¨ Detta innebar¨ oftast ett tidskravande¨ trial-and-error forfarande¨ for¨ att hitta lampliga¨ parametrar till ett specifikt DE-problem, framfor¨ allt for¨ utovare¨ med begransad¨ kunskap och erfarenhet av DE. Syftet med denna licentiatavhandling ar¨ att utveckla nya DE-algoritmer for¨ att behandla de ovannamnda¨ problemen. For¨ att mota¨ det forsta¨ problemet sa˚ studerades hybridisering av DE och lokala soktekniker¨ for¨ att effektivisera sokningen.¨ Tanken ar¨ att anvanda¨ en lokal sokmekanism¨ pa˚ den basta¨ indivi- den i varje generation i DE-algoritmen och utnyttja de mest lovande regionerna under evolutionsprocessen for¨ att snabba pa˚ konvergensen eller oka¨ chansen att undvika lokala optimum. Fyra lokala sokstrategier¨ har integrerats och testats i det globala DE-ramverket vilket har lett till fyra varianter av DE-algoritmerna med olika egenskaper betraffande¨ diversifiering och intensifiering. Till det an- dra problemet foresl¨ as˚ en greedy adaptation method for¨ dynamisk justering av kontrollparametrarna i DE. Den implementeras genom att utfora¨ greedy search upprepade ganger˚ under korningen¨ av DE for¨ att hitta battre¨ varden¨ till kon- trollparametrarna. Utvarderingen¨ av parameterval baseras pab˚ ade˚ success rate och fitness improvement av trial losningar¨ jamf¨ ort¨ med target losningar.¨ Sam- manslagningen av DE och denna greedy parameter adaptation har lett till en ny adaptiv DE-algoritm som kallas Greedy Adaptive Differential Evolution (GADE). Den foreslagna¨ metoden i denna licentiatavhandling har testats i olika pre-

iii iv standamatningar¨ och jamf¨ orts¨ med state-of-the-art-algoritmer, med goda resul- tat. Dessutom har den foreslagna¨ GADE-algoritmen anvants¨ i ett industriellt scenario och uppnadde˚ da˚ mer exakta resultat an¨ den med en standard DE- algoritm. To my family

Acknowledgements

First and foremost, I will like to express my sincere gratitude to my main su- pervisor Ning Xiong (Malardalen¨ University) and my two co-supervisors Pe- ter Funk (Malardalen¨ University) and Yigen Zenlander (ABB FACTS). I have learned a lot from their advices, feedback and guidance, without their support this thesis would not have been possible. I also would like to express the deep- est appreciation to my co-authors Magnus Evestedt (Prevas), Francisco Herrera (Granada University) and Daniel Molina (Cadiz University) for their coopera- tion in the papers. I would like to thank Baran C¸ ur¨ ukl¨ u¨ for helping me with the lectures and nice discussions. I would like to thank all my current and past roommates, Fredrik Ekstrand, Nikola Petrovic, Carl Ahlberg, Lennie Carlen´ Eriksson, Jonatan Tidare, Martin Ekstrom,¨ Gregory Koshmak, Jimmie Hagblad and Henrik Johansson. It is a pleasure to share office with you. I thank the lecturers and professors whom I learned a lot from during meet- ings, lectures, seminars and PhD courses including Thomas Nolte, Gordana Dodig-Crnkovic, Jan Gustafsson and Mikael Sjodin.¨ You all have been a source of inspiration for me. I thank the administrative staff in particular, Malin Ashuvud,˚ Sofia Jader¨ en,´ Carola Ryttersson and Jenny Hagglund¨ for making my live easier. I would like to thank my friends and colleagues at the division of IFT: Ankica, Branko, Elaine, Farid, Fredrik, Giacomo, Hakan,˚ Hamid, Hamidur, Ivan, Jiaying, Jonas, Jose-Fern´ an,´ Lars, Magnus, Maria Ehn, Maria Linden,´ Masoud, Mia, Miguel Angel, Mikael, Mirgita, Mobyen, Per, Per Olov, Sara, Shahina, Shaibal, Torbjorn,¨ Youssef and others. Also I would like to thank Radu, Inmaculada, Filip and others from other divisions. Last but not least, a huge thank to my family for their support even from the long distance. Also I would like to thank my girlfriend and best friend Laura, without your support this would be impossible. Miguel Leon, Vaster¨ as,˚ 2016

vii List of Publications

Papers included in the licentiate thesis1:

Paper A: Greedy Adaptation of Control Parameters in Differential Evo- lution for Problems. Miguel Leon, Ning Xiong, (CEC), 2015 IEEE Congress on, p. 385-392.

Paper B: Adaptive Differential Evolution Supports Automatic Model Cal- ibration in Furnace Optimized Control System. Miguel Leon, Mag- nus Evestedt, Ning Xiong, accepted in Studies in Computational Intelli- gence, Springer, 2016.

Paper C: Differential Evolution Enhanced with Eager Random Search for Solving Real-Parameter Optimization Problems. Miguel Leon, Ning Xiong. International Journal of Advanced Research in Artificial Intelli- gence, 2015, 1(4), p. 49-57.

Paper D: A new Differential Evolution Algorithm with Alopex-Based Lo- cal Search. Miguel Leon, Ning Xiong. International Conference on Artificial Intelligence and Soft Computing, 2016, p. 420-431.

1The included articles have been reformatted to comply with the licentiate layout

ix List of Publications

Papers included in the licentiate thesis1:

Paper A: Greedy Adaptation of Control Parameters in Differential Evo- lution for Global Optimization Problems. Miguel Leon, Ning Xiong, Evolutionary Computation (CEC), 2015 IEEE Congress on, p. 385-392.

Paper B: Adaptive Differential Evolution Supports Automatic Model Cal- ibration in Furnace Optimized Control System. Miguel Leon, Mag- nus Evestedt, Ning Xiong, accepted in Studies in Computational Intelli- gence, Springer, 2016.

Paper C: Differential Evolution Enhanced with Eager Random Search for Solving Real-Parameter Optimization Problems. Miguel Leon, Ning Xiong. International Journal of Advanced Research in Artificial Intelli- gence, 2015, 1(4), p. 49-57.

Paper D: A new Differential Evolution Algorithm with Alopex-Based Lo- cal Search. Miguel Leon, Ning Xiong. International Conference on Artificial Intelligence and Soft Computing, 2016, p. 420-431.

1The included articles have been reformatted to comply with the licentiate layout

ix x

Additional papers, not included in the licentiate thesis:

Investigation of mutation strategies in differential evolution for solv- • ing global optimization problems. Miguel Leon, Ning Xiong, Proc. on International Conference on Artificial Intelligence and Soft Computing, Springer, 2014, p. 372-383.

Using Random Local Search Helps in Avoiding Local Optimum in List of Figures • Differential Evolutions. Miguel Leon, Ning Xiong In Proc. IASTED, 2014, p. 372-383

Eager Random Search for Differential Evolution in Continuous Op- • 1.1 Domain-based evaluation ...... 5 timization. Miguel Leon, Ning Xiong, Progress in Artificial Intelli- gence, Springer, 2015, p. 286-291. 2.1 Differential Evolution process ...... 11 2.2 Random mutation with one difference vector ...... 11 Adapting Differential Evolution Algorithms For Continuous Opti- • 2.3 Mutation on the best individual with one difference vector . . 12 Miguel mization Via Greedy Adjustment Of Control Parameters. 2.4 current to best mutation with one difference vector ...... 13 Leon, Ning Xiong, Journal of Artificial Intelligence and Soft Computing 2.5 Example of binomial crossover ...... 13 Research, 6(2), 2016, p 103-118. 2.6 Example of exponential crossover (k=3,L=3) ...... 14 A Walk into for Engineering Optimization: Princi- • 7.1 DE, parameterized model, and temperature estimation . . . . . 61 ples, Methods and Recent Trends. Ning Xiong, Daniel Molina, Miguel 7.2 Random mutation with one difference vector ...... 62 Leon, Francisco Herrera, International Journal of Computational Intelli- 7.3 The evolution of the scaling factor F ...... 70 gence Systems 8(4), 2015, p. 606-636. 7.4 The evolution of the distribution center for the crossover prob- Principles and state-of-the-art of engineering optimization techniques. ability...... 71 • Ning Xiong, Miguel Leon, Proc. The 7th International Conference on 8.1 Random mutation strategy ...... 80 Advanced Engineering Computing and Applications in Sciences, 2013, 8.2 Current to best mutation strategy ...... 81 p. 36-42. 8.3 Current to rand mutation strategy ...... 82 8.4 Distribution probability ...... 86 8.5 Function 8 with two dimensions ...... 87 8.6 Function 9 with two dimensions ...... 88 8.7 Function 10 with two dimensions ...... 88

9.1 Random mutation with one difference vector ...... 104

xi List of Figures

1.1 Domain-based evaluation ...... 5

2.1 Differential Evolution process ...... 11 2.2 Random mutation with one difference vector ...... 11 2.3 Mutation on the best individual with one difference vector . . 12 2.4 current to best mutation with one difference vector ...... 13 2.5 Example of binomial crossover ...... 13 2.6 Example of exponential crossover (k=3,L=3) ...... 14

7.1 DE, parameterized model, and temperature estimation . . . . . 61 7.2 Random mutation with one difference vector ...... 62 7.3 The evolution of the scaling factor F ...... 70 7.4 The evolution of the distribution center for the crossover prob- ability...... 71

8.1 Random mutation strategy ...... 80 8.2 Current to best mutation strategy ...... 81 8.3 Current to rand mutation strategy ...... 82 8.4 Distribution probability ...... 86 8.5 Function 8 with two dimensions ...... 87 8.6 Function 9 with two dimensions ...... 88 8.7 Function 10 with two dimensions ...... 88

9.1 Random mutation with one difference vector ...... 104

xi

Abbreviations

ALS Alopex-based Local Search CR Crossover Rate DE Differential Evolution EA ERS Eager Random Search F mutation Factor GADE Greedy Adaptive Differential Evolution LIP Local Improvement Process LLS Local Improvement process oriented Local Search LS Local Search MA RC Research Contribution RQ Research Question XLS Crossover-based Local Search

xiii

Contents

I Thesis 1

1 Introduction 3 1.1ProblemFormulation...... 4 1.2ResearchQuestions...... 6 1.3 Scientific Contribution ...... 6 1.4ResearchMethod...... 8

2 Background 9 2.1AlopexSearch...... 9 2.2DifferentialEvolution...... 10

3 Related work 15 3.1 Parameter adaptation in Differential Evolution ...... 15 3.2 Local Search in Differential Evolution ...... 17 3.2.1 LLSinDE...... 17 3.2.2 XLSinDE...... 18 3.2.3 Strategies of Applying LS in DE ...... 18

4 Overview of the Papers 21

5 Conclusion and Future Work 25

Bibliography 27

II Included Papers 33

6 Paper A: Greedy Adaptation of Control Parameters in Differential Evolution

xv xvi Contents

for Global Optimization Problems 35 6.1Introduction...... 37 6.2BasicDE...... 38 6.3 Adaptation of Control Parameters in DE ...... 40 6.3.1 Parameter Adaptation in SaDE ...... 40 6.3.2 Parameter adaptation in JADE ...... 41 6.3.3 Parameter adaptation in MDE pBX...... 42 6.3.4 Parameter adaptation in SHADE ...... 42 6.4 Greedy Adaptation of DE Parameters ...... 43 6.4.1 Greedy Search for Parameter Adaptation ...... 44 6.4.2 Evaluation of Candidates for Parameter Assignment . 44 6.4.3 Greedy Adaptive DE Algorithm ...... 45 6.5Experiments...... 46 6.5.1 Experimental Settings ...... 46 6.5.2 Comparison of GADE with other algorithms . . . . . 49 6.6Conclusion...... 52

7 Paper B: Adaptive Differential Evolution Supports Automatic Model Cali- bration in Furnace Optimized Control System 57 7.1Introduction...... 59 7.2ProblemFormulation...... 60 7.3DifferentialEvolution...... 61 7.4 A New DE Algorithm with Parameter Adaptation ...... 63 7.4.1 Greedy Adjustment Scheme ...... 64 7.4.2 Adaptive DE Algorithm with Greedy Search . . . . . 66 7.5 Experiments and Results ...... 66 7.5.1 Experimental Settings ...... 68 7.5.2 Results and Comparison in Problem 1 ...... 68 7.5.3 Results and Comparison in Problem 2 ...... 69 7.5.4 Evolution of the DE Parameters ...... 70 7.6RelevantWork...... 71 7.7Conclusion...... 72

8 Paper C: Differential Evolution Enhanced with Eager Random Search for Solving Real-Parameter Optimization Problems 75 8.1Introduction...... 77 8.2RelatedWork...... 78 8.3BasicDE...... 79 Contents xvii

8.4DEIntegratedwithERS...... 83 8.4.1 Eager Random Local Search (ERS) ...... 83 8.4.2 The Proposed Memetic DE Algorithm ...... 86 8.5 Experiments and Results ...... 86 8.5.1 Experimental Settings ...... 89 8.5.2 Performance of the Memetic DE with random mutation strategy...... 90 8.5.3 Performance of the Memetic DE with Current to Rand MutationStrategy...... 90 8.5.4 Performance of the Memetic DE with Current to Best MutationStrategy...... 92 8.6Conclusions...... 94

9 Paper D: A new Differential Evolution Algorithm with Alopex-Based Local Search 99 9.1Introduction...... 101 9.2RelatedWork...... 102 9.3 Differential Evolution ...... 103 9.4 Differential Evolution with Alopex Local Search ...... 104 9.4.1 Alopex...... 105 9.4.2 DEALS...... 106 9.5 Experiments and Results ...... 107 9.5.1 Experimental Settings ...... 107 9.5.2 Comparing DEALS with Basic DE and DENLS . . . . 108 9.5.3 Comparing DEALS with others algorithms ...... 109 9.6Conclusion...... 110

I

Thesis

1 Chapter 1

Introduction

Nowadays, optimization is one of the most important issues in industrial sys- tems design and product development [1, 2, 3, 4]. Optimization algorithms enhance the performance of industrial products while they help reducing the cost to be more competitive. In the industrial world, companies often need to optimize one or more factors (less cost, better performance, etc.), where opti- mization can play an important role, adjusting or fine tuning system designs in terms of the factors mentioned before. This is not a trivial task, especially when the problem space is complex and has high dimensionality. Metaheuristics [5, 6, 7] are promising techniques that are beneficially used when classic optimization algorithms cannot solve complex a problems effec- tively. Metaheuristics can solve a wide range of problems with different char- acteristics and they are able to find good solutions, acceptable solutions, in a feasible time, compared with traditional algorithms. Metaheuristics for opti- mization can be divided into two categories: trajectory-based approaches and population-based approaches. Trajectory-based algorithms make transitions from a possible solution to a different one (sometimes even worse solutions are also considered for a transition). Some examples of trajectory-based ap- proaches are hill-climbing [5], [8], [9], gra- dient descent [10] or simplex method [11]. Multi-start strategy has also been used here to obtain better global solutions, leading to multi-trajectory based al- gorithms such as iterative local search [6], variable neighborhood search [12], greedy randomized adaptive search procedure [13] and iterative greedy search [14]. In population-based approaches such as evolutionary algorithms (EAs), a group of solutions (also called population) are handled at the same time, with a new population being generated from an old one. During the last decades, EAs have been proved as a powerful tool to solve many real-world optimization problems. They can find a good solution (probably not the theoretically opti-

3 Chapter 1

Introduction

Nowadays, optimization is one of the most important issues in industrial sys- tems design and product development [1, 2, 3, 4]. Optimization algorithms enhance the performance of industrial products while they help reducing the cost to be more competitive. In the industrial world, companies often need to optimize one or more factors (less cost, better performance, etc.), where opti- mization can play an important role, adjusting or fine tuning system designs in terms of the factors mentioned before. This is not a trivial task, especially when the problem space is complex and has high dimensionality. Metaheuristics [5, 6, 7] are promising techniques that are beneficially used when classic optimization algorithms cannot solve complex a problems effec- tively. Metaheuristics can solve a wide range of problems with different char- acteristics and they are able to find good solutions, acceptable solutions, in a feasible time, compared with traditional algorithms. Metaheuristics for opti- mization can be divided into two categories: trajectory-based approaches and population-based approaches. Trajectory-based algorithms make transitions from a possible solution to a different one (sometimes even worse solutions are also considered for a transition). Some examples of trajectory-based ap- proaches are hill-climbing [5], simulated annealing [8], tabu search [9], gra- dient descent [10] or simplex method [11]. Multi-start strategy has also been used here to obtain better global solutions, leading to multi-trajectory based al- gorithms such as iterative local search [6], variable neighborhood search [12], greedy randomized adaptive search procedure [13] and iterative greedy search [14]. In population-based approaches such as evolutionary algorithms (EAs), a group of solutions (also called population) are handled at the same time, with a new population being generated from an old one. During the last decades, EAs have been proved as a powerful tool to solve many real-world optimization problems. They can find a good solution (probably not the theoretically opti-

3 4 Chapter 1. Introduction 1.1 Problem Formulation 5 mal one) of a NP-hard problem in a feasible time. There is a big family of EAs, All points that satisfy the condition in Eq. (1.1) are termed as stationary but the most famous are Differential Evolution (DE) [7], points. Further, in order to be a minimum solution, the of the (GA) [15] and Particle Swarm Optimization (PSO) [16]. Other algorithms can second derivatives, as defined in Eq. (1.2), has to be positive definite at the be found in [17]. stationary point. Hybridization of EAs with local search (LS) algorithms [18, 19] has been a hot research topic. Applying local search techniques can bring two benefits to ∂2f ∂2f ∂2f the whole search procedure. First, they can drive the search into a better area 2 ... ∂x1 ∂x1∂x2 ∂x1∂xn further from local optima. Second, but not less important, they can enhance ∂2f ∂2f ∂2f ∂x ∂x 2 ... ∂x ∂x H = 2 1 ∂x2 2 n . (1.2) the exploitation of some promising areas of the search space so as to speed ...... up the convergence of the search. These hybridized algorithms are also called ∂2f ∂2f ∂2f ... 2 Memetic Algorithms (MAs) [20, 21] in the literature. ∂xn∂x1 ∂xn∂x2 ∂xn Differential Evolution (DE) has been a competitive class of EAs. It is char- acterized by the way in which new solutions are created based on differences However, when it comes to an engineering problem, normally an analytic between pair(s) of individuals randomly selected from the population. DE has formulation of the objective function is not available, not to mention the sec- been popular in many applications due to its high ability, simple implementa- ond order partial derivatives. Very frequently the objective value for a possible tion and easy use with a relatively low number of control parameters [4, 22]. solution is derived from a specific domain-based evaluation such as simulation, However, the performance of DE is not always excellent. On one hand, DE can as shown in Fig. 1.1. Hence we need to seek numerically approximate solu- occasionally get stuck in local optima causing unsatisfactory final solutions. tions instead of mathematically exact solutions in such practical engineering On the other hand, DE has some operations and parameters that are problem scenarios. dependent. This means that, depending on the characteristics of distinct prob- This thesis focuses on Differential Evolution as a class of lems, different parameter values will be required. Improper setting of control techniques to search for approximate solutions to real-parameter optimization parameters for DE will lead to poor optimization performance. tasks. The goal is to develop new methods and algorithms to enhance standard This thesis aims to develop new methods and techniques to mitigate the two DE algorithm in dealing with a wide variety of complex problems. aforementioned problems of DE in solving complex optimization problems. First, suitable local search schemes have been investigated and incorporated into DE to reduce the risk of local stagnation. Second, adaptive DE algorithm has been developed allowing for automatic adjustment of control parameters of DE without requiring prior knowledge of the problem.

1.1 Problem Formulation

The problem of optimization, without loss of generality, can be formulated as the task of finding a vector of variables X =(x1,x2,...,xn) inside the search space Ω, to minimize the objective function f . Therefore, no any other point in the search space will have an objective value larger than f(X∗). In mathematical principle, an optimal solution to a nonlinear objective func- tion must have all its partial derivatives equal to zero, as shown in Eq. (1.1),

∂f Figure 1.1: Domain-based evaluation =0,i=1, 2,...,n (1.1) ∂xi 1.1 Problem Formulation 5

All points that satisfy the condition in Eq. (1.1) are termed as stationary points. Further, in order to be a minimum solution, the Hessian matrix of the second derivatives, as defined in Eq. (1.2), has to be positive definite at the stationary point.

∂2f ∂2f ∂2f 2 ... ∂x1 ∂x1∂x2 ∂x1∂xn ∂2f ∂2f ∂2f ∂x ∂x 2 ... ∂x ∂x H = 2 1 ∂x2 2 n . (1.2) ...... ∂2f ∂2f ∂2f ... 2 ∂xn∂x1 ∂xn∂x2 ∂xn However, when it comes to an engineering problem, normally an analytic formulation of the objective function is not available, not to mention the sec- ond order partial derivatives. Very frequently the objective value for a possible solution is derived from a specific domain-based evaluation such as simulation, as shown in Fig. 1.1. Hence we need to seek numerically approximate solu- tions instead of mathematically exact solutions in such practical engineering scenarios. This thesis focuses on Differential Evolution as a class of metaheuristic techniques to search for approximate solutions to real-parameter optimization tasks. The goal is to develop new methods and algorithms to enhance standard DE algorithm in dealing with a wide variety of complex problems.

Figure 1.1: Domain-based evaluation 6 Chapter 1. Introduction

1.2 Research Questions

Given the weakness of DE as stated before, it is important to address the fol- lowing two research questions to further augment the performance of DE in wide engineering applications.

RQ1: How an online adaptation of the main parameters of DE can be made to make search more efficient in different regions of the space with different properties? Mutation factor and crossover rate are the main control parameters that affect the search process of differential evolution. Depending on the mu- tation factor value, the movement step size inside DE can be larger or smaller, therefore this value should be determined based on the state in which the evolutionary process is situated. The crossover rate affects the number of variables that will be altered in every iteration of the evolu- tionary process, hence its setting will be dependent on the characteristics of the problem at hand

RQ2: How local search can be conducted inside the global search process of DE to enhance its performance whereas requiring no derivative informa- tion? It is well known that descent can be used as a local search method to locate new positions to quickly reduce objective values. How- ever, approximating derivative information requires evaluations of a num- ber of adjacent points, which would consume too much computing re- source. An interesting issue for investigation is how DE could be com- bined with derivative-free local search for improving its exploration and exploitation capability, leading to higher quality solutions found in shorter time.

1.3 Scientific Contribution

In this section, we present the contributions of the thesis that address the previ- ously formulated research questions.

SC1: The first research contribution of this thesis deals with RQ1, providing a method to adapt mutation factor and crossover rate inside DE. An adap- tive differential evolution algorithm called “Greedy Adaptive Differential Evolution” (GADE) has been developed in paper A. The adaptation ap- proach proposed for GADE can automatically adjust mutation factor and 1.3 Scientific Contribution 7

crossover rate during the course of search while solving a specific prob- lem. It relies on experience to update both parameters online , adopting a new method to calculate the progress rate that a candidate (for muta- tion factor and crossover rate) has accomplished. In paper B, GADE was further tested and verified with the application to a practical industrial scenario.

SC2: The second scientific contribution tackles RQ2, proposing the possibili- ties of incorporating four derivative-free local search schemes into a stan- dard DE algorithm. Three of these approaches were handled in Paper C and they belong to a group called “Eager Local Search”. Each method in that group uses a different probability distribution, in order to create random perturbation to the best individual in the population, as is sum- marized below:

Random Local Search uses a uniform distribution in order to create • evenly distributed perturbations when testing new positions in the search space. Cauchy Local Search uses a Cauchy distribution to perform a small • perturbation with high probability, but still it gives low chance to perform a large perturbation. Normal Local Search uses a normal distribution to offer the high- • est chance of small perturbation in comparison with uniform and Cauchy distributions.

The forth method is called “Alopex-based Local Search” (ALS) and it was investigated in paper D. This approach, using a temperature pa- rameter that changes with state of the evolutionary process, makes large moves at the early stage of the search and small moves at the late phase. By means of this, the balance between exploration and exploitation is adapted automatically.

Papers A, B, C, and D are included in the thesis. The relation between the included papers and the research questions is given in Table 1. 8 Chapter 1. Introduction

Table 1.1: Relation between the included papers and research questions paper RQ1 RQ2 PAPER A X PAPER B X PAPER C X PAPER D X

1.4 Research Method

In this thesis, the basic research methodology was to identify the research prob- lems by studying the state-of-art and related works and then attempt to propose new methods and algorithms to solve the problems. For evaluation, we em- ployed an empirical method to evaluate the proposed solutions mainly on a set of benchmark problems that are widely used by the research community. Essentially the research has been conducted iteratively in the following steps:

1. Perform literature study to identify research problems

2. Propose a solution (algorithm or method) to solve the identified research problems

3. Implement the proposed solution in a software system

4. Evaluate the proposed solution in different benchmark problems given by the research community. It is also possible to do evaluation in practical industrial problems.

5. Compare the results of the proposed solution with the current state-of-art algorithms Chapter 2

Background

Numerical methods for optimization aim to create progressively improved ap- proximation to the problem in a number of iterations [17]. Generally, they can be divided into two large categories: trajectory-based approaches and population- based approaches. In trajectory-based approaches only one new solution is generated in each iteration, while population-based approaches produce a new population of solutions from the preceding population. In relevance to the re- search of the thesis, this chapter will introduce Alopex [23] search that belongs to the first category (trajectory-based) and standard Differential Evolution (DE) algorithm that belongs to the second category (population-based).

2.1 Alopex Search

Harth and Tzanakou proposed Alopex (Algorithm of pattern extraction) for optimization and pattern in visual receptive fields [23]. Alopex aims to estimate the gradient of the objective function by measuring the changes between two independent solutions. In the first step, the correlation between two solutions, A =(a1,a2,...,an) and B =(b1,b2,...,bn), is calculated following Eq. (2.1).

C =(a b ) [f(A) f(B)] for j =1, 2,...,n (2.1) j j − j · − where f(A) and f(B) are the objective values of A and B, respectively. Considering a minimization problem as an example, the probability of a negative move on the jth variable is calculated using the Bolzmann distribution as

9 10 Chapter 2. Background 2.2 Differential Evolution 11

1 Pj = C (2.2) − j 1+e T where T is obtained following Eq. (2.3), in which the average of the absolute value from all the correlations of all variables is calculated.

n 1 T = C (2.3) n · | j| j=1 The direction of the move is represented by δj and it is determined accord- ing Eq. (2.4).

1 if Pj rand(0, 1) Figure 2.1: Differential Evolution process δj = − ≥ (2.4) 1 otherwise

In the final step, the new trial solution, Q =(q1,q2,...,qn), is calculated The first mutation strategy is random mutation strategy, in which three ran- as domly selected individuals from the population are used to generate the mutant vector. More precisely, this mutant vector is created according to Eq. (2.6). q = a + δ a b N(0,λ) (2.5) j j j ·| j − j|· Vi,g = Xr1,g + F (Xr2,g Xr3,g) (2.6) where N(0,λ) is a random value between and following a normal · − −∞ ∞ distribution with mean equal to 0 and standard deviation equal to λ. If f(Q) is where Vi,g represents the mutant vector, i stands for the index of the vector, g stands for the generation, and r ,r ,r 1, 2,...,Np are random integers. better or equal than f(A), then A is replaced by T. This process can continue 1 2 3 ∈{ } until a stop criteria is satisfied, i.e., the maximum number of evaluations have F is the scaling factor in the interval [0, 2]. Fig. 2.2 shows how this mutation been reached or a sufficiently good solution has been found. strategy works in a two dimensional problem. All the variables in the figure appear in Eq. (2.6) with the same meaning, and d stands for the difference

vector between the individuals Xr2,g and Xr3,g. This mutation method is often 2.2 Differential Evolution represented by DE/rand/1, because it only uses one difference vector.

DE is a stochastic and population-based algorithm. The population is composed of Np individuals and every individual in the population represents a possible solution to the problem. DE operates in three consecutive steps in every iter- ation: mutation, crossover and selection, as shown in the diagram in Fig. 2.1. The explanation of these steps is given below: MUTATION. Np mutated individuals are generated using some individuals of the population. A vector for the mutated solution is called mutant vector. There are different strategies to create a mutant vector. Usually they are notated as DE/x/y/, where x stands for the vector to be mutated and y represents the number of difference vectors used in the mutation. Here only the three most common mutation methods are explained. Other mutation strategies and their performance have been discussed in [24]. Figure 2.2: Random mutation with one difference vector 2.2 Differential Evolution 11

Figure 2.1: Differential Evolution process

The first mutation strategy is random mutation strategy, in which three ran- domly selected individuals from the population are used to generate the mutant vector. More precisely, this mutant vector is created according to Eq. (2.6).

V = X + F (X X ) (2.6) i,g r1,g · r2,g − r3,g where Vi,g represents the mutant vector, i stands for the index of the vector, g stands for the generation, and r ,r ,r 1, 2,...,Np are random integers. 1 2 3 ∈{ } F is the scaling factor in the interval [0, 2]. Fig. 2.2 shows how this mutation strategy works in a two dimensional problem. All the variables in the figure appear in Eq. (2.6) with the same meaning, and d stands for the difference vector between the individuals Xr2,g and Xr3,g. This mutation method is often represented by DE/rand/1, because it only uses one difference vector.

Figure 2.2: Random mutation with one difference vector 12 Chapter 2. Background 2.2 Differential Evolution 13

The second mutation strategy uses the best individual from the population to create the mutant vectors. This strategy is represented by DE/best/1 and the mutant vectors are created in terms of Eq. (2.7).

V = X + F (X X ) (2.7) i,g best,g · r1,g − r2,g where Xbest,g represents the best individual in the population at generation g. Fig. 2.3 shows how DE/best/1 works in a two dimensional problem. All the variables in the figure appear in Eq. (2.7) with the same meaning, and d stands for the vector of difference between the individuals Xr1,g and Xr2,g

Figure 2.4: current to best mutation with one difference vector

index and g refers to the generation. There are two basic crossover strategies [25]: binomial and exponential. The binomial crossover employs Eq. (2.9) to create an offspring solution.

Vi,g[j] if rand[0, 1] < CR or j = jrand Ti,g[j]= (2.9) Xi,g[j] otherwise where j stands for the index of every parameter in a vector, CR is the probability Figure 2.3: Mutation on the best individual with one difference vector of the recombination and jrand is a randomly selected integer between 1 and Np to ensure that the offspring will not be the same as the parent. In binomial The third mutation strategy moves the current individual towards the best crossover, the inherited parts from the mutant vector are randomly distributed individual in the population before being disturbed with a scaled difference across all dimensions. An example is given in Fig. 2.5. of two randomly selected individuals. This mutation strategy is notated as DE/current-to-best/1. Thus, the mutant individual is created according to Eq. (2.8).

V = X + F 1 (X X )+F 2 (X X ) (2.8) i,g i,g · best,g − i,g · r1,g − r2,g where Xi,g represents the current individual at generation g. Fig. 2.4 shows how DE/current-to-best/1 works in a two dimensional problem. All the vari- ables in the figure appear in Eq. (2.8) with the same meaning, and d1 stands for the difference vector between the current individual Xi,g and the best individual Figure 2.5: Example of binomial crossover

Xbest,g, and d2 denotes the difference vector between Xr1,g and Xr2,g. CROSSOVER.In the second step, we recombine the set of mutant vectors The exponential crossover constructs the offspring vector containing a se- created in the first stage (mutation) with the original population members to quence of consecutive components (in a circular manner) taken from the mutant generate offspring solutions. A new offspring is denoted by Ti,g where i is the vector. The offspring vector is derived, in Eq. (2.10), as follows 2.2 Differential Evolution 13

Figure 2.4: current to best mutation with one difference vector index and g refers to the generation. There are two basic crossover strategies [25]: binomial and exponential. The binomial crossover employs Eq. (2.9) to create an offspring solution.

Vi,g[j] if rand[0, 1] < CR or j = jrand Ti,g[j]= (2.9) Xi,g[j] otherwise where j stands for the index of every parameter in a vector, CR is the probability of the recombination and jrand is a randomly selected integer between 1 and Np to ensure that the offspring will not be the same as the parent. In binomial crossover, the inherited parts from the mutant vector are randomly distributed across all dimensions. An example is given in Fig. 2.5.

Figure 2.5: Example of binomial crossover

The exponential crossover constructs the offspring vector containing a se- quence of consecutive components (in a circular manner) taken from the mutant vector. The offspring vector is derived, in Eq. (2.10), as follows 14 Chapter 2. Background

Vi,g[j] if j k, k +1 n, k +2 n,..., k + L 1 n Ti,g[j]= ∈{ { } { } { − } } Xi,g[j] otherwise (2.10) where k and L are two randomly selected integers in 1, 2,...,(n 1) , n { − } stands for the dimension of the problem and L represents the number of ele- Chapter 3 ments that will be taken from the mutant vector. x is x if x n and x n { }n ≤ − if x>n. An example (with k=3, L=3) is given in Fig. 2.6. Related work

This chapter presents a survey of the related works. Section 3.1 presents an overview of the previous works on adaptions of the main parameters of DE. Section 3.2 outlines different local search methods that have been used in co- operation with Differential Evolution.

Figure 2.6: Example of exponential crossover (k=3,L=3) 3.1 Parameter adaptation in Differential Evolution

SELECTION. In this last step, the offspring is compared with its parent, Researchers have proposed different methods to adapt the main control param- and the stronger one will enter the population for the next generation. There- eters: mutation factor (F), crossover rate (CR) and population size (NP) for fore, if the problem to solve is a minimization problem, the next generation is Differential Evolution. Qin and Suganthan proposed the SaDE algorithm [26], created according to Eq. (2.11). as one of the earliest works to adapt parameters in DE. Their algorithm uses normal distributions to assign mutation factor and crossover rate to an indi- Ti,g if f(Ti,g)

15 Chapter 3

Related work

This chapter presents a survey of the related works. Section 3.1 presents an overview of the previous works on adaptions of the main parameters of DE. Section 3.2 outlines different local search methods that have been used in co- operation with Differential Evolution.

3.1 Parameter adaptation in Differential Evolution

Researchers have proposed different methods to adapt the main control param- eters: mutation factor (F), crossover rate (CR) and population size (NP) for Differential Evolution. Qin and Suganthan proposed the SaDE algorithm [26], as one of the earliest works to adapt parameters in DE. Their algorithm uses normal distributions to assign mutation factor and crossover rate to an indi- vidual in the population. SaDE also maintains a memory of the successful CR values, the median of them is used as the mean of the distribution for CR. Other works that are relevant to SaDE can be found in [27, 28]. The idea of using successful values (for F and CR) in order to change the mean of a distribution has been commonly practiced by many researchers. In the JADE algorithm [29] proposed by Zhang and Sanderson, Cauchy distri- bution and normal distribution are used to determine the mutation factor and the crossover rate, respectively. The mean values of both distributions are up- dated after each generation using Lehmer mean and arithmetic mean of the successful parameter values respectively. Similarly Islam and Das proposed the MDE pBX algorithm [30], which uses the same distributions but a different mean (called pow) to update the centres of both distributions. Likewise Wang proposed the dn DADE algorithm[31], in which the the mean and the stan- dard deviation of the normal distributions are updated. An improved version of

15 16 Chapter 3. Related work

JADE was given in [32] in which the standard deviation of the distributions is modified in a similar way to that in dn DADE.

More recently Tanabe and Fukunaga proposed the SHADE algorithm [33], in which the same distributions as in JADE are used to generate control param- eters. The difference between both algorithms is that JADE uses a single pair of mean values to guide the parameter adaptation, while in SHADE a historical record of successful parameter settings are updated in a cycle and also used during the evolutionary process.

A different method termed as EPSDE was proposed by Mallipedi and Sug- anthan [34]. Therein the mutation factor and crossover rate are selected from different pools of values. The pool of F values is taken in the range 0.4-0.9 with steps of 0.1, and the pool of CR values is taken in the range 0.1-0.9 in steps of 0.1. Successful combinations of F and CR values are stored and they will be used in future generations. If the combination for an individual is not successful, it will be reinitialized with a new random combination or with a combination randomly selected from the storage of successful combinations. Similar work was conducted by Wang et al. [35] in the development of the CoDE method, which only uses one pool containing three different combina- tions of F and CR values. Brest et al. proposed the jDE algorithm [36] for self- adapting control parameters in DE. The key to this algorithm is the extension of individuals in the population with parameter values. Thus the adjustment of control parameters is implemented by evolution on the individual level. Gen- erally, better parameter values result in new individuals of higher quality and thereby acquiring more chances to propagate themselves.

Liu and Lampinen proposed a fuzzy adaptive differential evolution algo- rithm (FADE) [37]. This method adapts the control parameters (F and CR) of DE employing fuzzy logic controllers whose inputs are the relative objective function values and individuals of successive generations.

Besides, population size adaptation has also received some attention from researchers. In [38], Teo proposed two versions of an algorithm for self-adapting populations (DESAP-Rel and DESAP-Abs). In DESAP-Abs the population size of a subsequent generation is set as the average of the population size at- tribute for all members of the current population. However, DESAP-Rel takes the current population size plus an increasing/decreasing percentage in terms of the population growth rate. 3.2 Local Search in Differential Evolution 17

3.2 Local Search in Differential Evolution

Incorporating a local search (LS) heuristic is often very useful in the design of an effective evolutionary algorithm for global optimization. Local search techniques can be divided into two groups according to Lozano et al. [39]:

Local improvement process (LIP) oriented LS (LLS): Local search heuris- • tics in this group applies local improvements on one or more individuals in the population. Some well known techniques, like hill-climbing, gra- dient descent, or some random or predefined perturbations can be used here. LIPs can be applied after a certain number of generations and pos- sibly with some probability of doing it.

Crossover-based LS (XLS): In this second group, the local search is im- • plemented by recombination operators, in which the individuals from the population are used as parents to produce offspring around them. These techniques have an attractive property of self-adaptation [40], since the population is evolving as the search progresses. As in LIPs, XLS can be applied after a certain number of generations.

Next, we shall review the previous works of LS in these two classes that are incorporated into a global DE algorithm.

3.2.1 LLS in DE There are many works that belong to LLS. One of the ideas is to use the chaos theory [41] in order to create random perturbations for local search in DE. Pei- chong et al. [42] proposed DECH algorithm, a hybridization of DE with a chaos local search (CLS) approach. In this algorithm, CLS is applied to the best 20% of the individuals after a regular DE iteration .If the perturbation makes the new individual better, it will enter the population as a member of the next generation. The other individuals are discarded and then regenerated randomly. A similar work was performed by Jia et al. [43], where CLS is used after every generation on the best individual. The local search in [44] performs a perturbation p to the individual on every single dimension. This perturbation first is done in one direction and only if it does not improve, then the perturbation will be performed in the opposite direction. If the final solution does not improve after trying on all dimensions, the current step size p will be halved for the next iteration. This local search method uses a large number of function evaluations, so a local limitation has been made to avoid over usage of computing resource by LS. 18 Chapter 3. Related work

3.2.2 XLS in DE

Crossover based local search methods have also attained much research atten- tion. Noman and Iba [45] proposed the algorithm called DEfirSPX, in which a simplex crossover (SPX) is used. The best individual and (p-1) random individ- uals are selected. SPX is applied with a number of times to these individuals, and then the best solutions obtained will participate in the selection process in- side the DE procedure. An improved version of this algorithm was proposed in [46], which is called DEahcSPX. Different from DEfirSPX which uses a fixed search length, the search length is adapted in DEahcSPX using an adaptive strategy. This strategy makes use of feedback from the search in order to select a suitable length of search in the dynamic process. Gu and Gu [47] proposed a local search method that uses difference be- tween the current individual and the best individual in the population to mod- ify the current individual. After every generation this LS operator is applied with a probability to all individuals in the population. Alternatively, two other approaches were proposed in [48]. The first method, called DETLS, uses a trigonometric local search, in which a weighted average of three points from the population (the best individual and two other random individuals) is cal- culated as the new individual. The second method, called DEILS, is based on quadratic interpolation. As in DETLS, the same three individuals are selected, but in this case a parabolic curve is created and then the minimum point in the curve is identified. In both algorithms (DEILS and DETLS) the local search is applied after every generation and it continues until no further improvement. Peng and Wu [49] proposed a heterozygous differential evolution with Taguchi local search method (THDE). Taguchi method uses two random individuals from the population and the middle point between them (each parent is divided into four parts or factors) in order to create nine new individuals. The effect of the three parents is calculated and for each factor, the part of the best parent is copied to the offspring. Subsequently a random individual in the popula- tion is replaced by the resultant offspring. A similar approach proposed in [19] focuses on the worst individual in the population, which is replaced by the re- sultant offspring if it is weaker than the offspring.

3.2.3 Strategies of Applying LS in DE

Researches follow different strategies when they apply a LS technique in DE. In the methods such as DECH [42], DECLS [43], DETLS and DEILS [48] as explained above, local search is applied after every generation of DE. Different from these methods, there are some hybridizations that perform LS after a cer- 3.2 Local Search in Differential Evolution 19 tain number of generations. Two different approaches, that follow this strategy, were proposed in [50], in which local search chains (LSC) [20] are applied to SaDE [26]. The first method, SaDELAM , updates the population after perform- ing the LSC. In the second method (SaDEBAL), the population is not updated after LCS is applied, but the new created individuals are saved in an external archive. In order to apply LSC again, the best individual from the external archive and the population will be used as the starting point. This strategy is also used in [44]. Another key issue to take into account is to which individuals of the pop- ulation the local search technique should be applied. In some algorithms, the LS technique is only applied to the best individual ins the population [43, 48]. Applying LS to the best individual is the most commonly adopted option by the researchers, but others alternatives can also be used. For instance, in [42], LS is applied on a percentage of the best individuals in the population. Alternatively, in [44], LS is applied to the individuals with their fitness values higher than the average. In another option presented in [47], LS is applied with a probability to all individuals of the complete population.

Chapter 4

Overview of the Papers

Paper A: Greedy adaptation of control parameters in differential • evolution for global optimization problems: Miguel Leon, Ning Xiong, Evolutionary Computation (CEC), 2015 IEEE Congress on, p. 385-392.

Abstract: Differential evolution (DE) is a very attractive evolutionary and meta-heuristic technique to solve many optimization problems in various real-world scenarios. However, the proper setting of con- trol parameters of DE is highly dependent on the problem to solve as well as on the different stages of the search process. This pa- per proposes a new greedy adaptation method for dynamic adjust- ment of mutation factor and crossover rate in DE. The proposed method is based on the idea of greedy search to find better param- eter assignment in the neighborhood of a current candidate. Our work emphasizes reliable evaluation of candidates via applying a candidate with a number of times in the search process. As our pur- pose is not merely to increase the success rate (the survival of more trial solutions) but also to accelerate the speed of fitness improve- ment, we suggest a new metric termed as progress rate to access the quality of candidates in support of the greedy search. This greedy parameter adaptation method has been incorporated into basic DE, leading to a new DE algorithm called Greedy Adaptive Differential Evolution (GADE). GADE was tested on 25 benchmark functions in comparison with five other DE variants. The results of evalua- tion demonstrate that GADE is strongly competitive: it obtains the best ranking among the counterparts in terms of the summation of relative errors across the benchmark functions. My Contribution: I am the main author of the paper contributing with

21 22 Chapter 4. Overview of the Papers

the idea, the algorithm, the implementation and the experiments. Also, I wrote the major draft of the paper.

Paper B: Adaptive Differential Evolution Supports Automatic Model • Calibration in Furnace Optimized Control System: Miguel Leon, Magnus Evestedt, Ning Xiong, accepted in Studies in Computational In- telligence, Springer, 2016

Abstract: Model calibration represents the task of estimating the pa- rameters of a process model to obtain a good match between ob- served and simulated behaviours. This can be considered as an op- timization problem to search for model parameters that minimize the discrepancy between the model outputs and the corresponding features from the historical empirical data. This chapter investi- gates the use of Differential Evolution (DE), a competitive class of evolutionary algorithms, to solve calibration problems for nonlin- ear process models. The merits of DE include simple and compact structure, easy implementation, as well as high convergence speed. However, the good performance of DE relies on proper setting of its running parameters such as scaling factor and crossover proba- bility, which are problem dependent and which can even vary in the different stages of the search. To mitigate this issue, we propose a new adaptive DE algorithm that dynamically adjusts its running parameters during its execution. The core of this new algorithm is the incorporated greedy local search, which is conducted in suc- cessive learning periods to continuously locate better parameter as- signments in the optimization process. In case studies, we have applied our proposed adaptive DE algorithm for model calibration in a Furnace Optimized Control System. The statistical analysis of experimental results demonstrate that the proposed DE algorithm can support the creation of process models that are more accurate than those produced by standard DE. My Contribution: I am the main author of the paper contributing with the idea, the algorithm, the implementation and the experiments. Also, I wrote the major draft of the paper.

Paper C: Differential Evolution Enhanced with Eager Random Search • for Solving Real-Parameter Optimization Problems: Miguel Leon, Ning Xiong, International Journal of Advanced Research in Artificial In- telligence, 2015, 1(4), p. 49-57 23

Abstract: Differential evolution (DE) presents a class of evolutionary computing techniques that appear effective to handle real parame- ter optimization tasks in many practical applications. However, the performance of DE is not always perfect to ensure fast convergence to the global optimum. It can easily get stagnation resulting in low precision of acquired results or even failure. This paper proposes a new memetic DE algorithm by incorporating Eager Random Search (ERS) to enhance the performance of a basic DE algorithm. ERS is a local search method that is eager to replace the current solution by a better candidate in the neighborhood. Three concrete local search strategies for ERS are further introduced and discussed, leading to variants of the proposed memetic DE algorithm. In addition, only a small subset of randomly selected variables is used in each step of the local search for randomly deciding the next trial solution. The results of tests on a set of benchmark problems have demon- strated that the hybridization of DE with Eager Random Search can substantially augment DE algorithms to find better or more precise solutions while not requiring extra computing resources. My Contribution: I am the main author of the paper contributing with the idea, the algorithm, the implementation and the experiments. Also, I wrote the major draft of the paper.

Paper D: A New Differential Evolution Algorithm with Alopex-Based • Local Search: Miguel Leon, Ning Xiong, Proc. on International Con- ference on Artificial Intelligence and Soft Computing, Springer, 2016.

Abstract: Differential evolution (DE), as a class of biologically inspired and meta-heuristic techniques, has attained increasing popularity in solving many real world optimization problems. However, DE is not always successful. It can easily get stuck in a local optimum or an undesired stagnation condition. This paper proposes a new DE algorithm Differential Evolution with Alopex-Based Local Search (DEALS), for enhancing DE performance. Alopex uses local cor- relations between changes in individual parameters and changes in function values to estimate the gradient of the landscape. It also contains the idea of simulated annealing that uses temperature to control the probability of move directions during the search pro- cess. The results from experiments demonstrate that the use of Alopex as local search in DE brings substantial performance im- provement over the standard DE algorithm. The proposed DEALS 24 Chapter 4. Overview of the Papers

algorithm has also been shown to be strongly competitive (best rank) against several other DE variants with local search. My Contribution: I am the main author of the paper contributing with the idea, the algorithm, the implementation and the experiments. Also, I wrote the major draft of the paper. Chapter 5

Conclusion and Future Work

In this thesis, we have proposed different methods to enhance the search power of DE. First, the Greedy Adaptive Differential Evolution (GADE) algorithm has been developed. GADE is able to adapt its mutation factor and crossover rate online using the greedy search scheme as well as a new metric on progress rate to evaluate the candidates of parameter assignment. The results of the experiments demonstrate that the adaptation of the parameters improves the performance of DE and that our proposed GADE algorithm is rather compet- itive to the state-of-the-art algorithms. Additionally, GADE has been applied and tested in a practical industrial problem with good results that outperform the standard DE algorithm. GADE should be attractive in engineering applica- tions, since it is easy to use without requiring prior knowledge of DE. Secondly, four different LS strategies have been investigated when being incorporated into DE. Three of them can be encapsulated into a general group called Eager Random Search (ERS). The three approaches in ERS share the same principle in that random perturbations are made to the individuals in order to improve the performance of solutions. The difference among the three lies in the different distributions used. Using normal or Cauchy distribution leads to discovery of better results in unimodal functions, since the perturbations made to individuals are more likely to be small. Using a uniform distribution is expected to give good results in multimodal functions containing multiple local optima. The forth proposed LS technique uses the Alopex (ALS) method in order to enable big moves at the beginning and small moves at the late stages of the search. Hence it can automatically balance exploration and exploitation with the evolutionary process. DEALS is shown to outperform some other memetic algorithms (MAs) in a set of benchmark problems. So far the research in the thesis employs a single population for DE. A multi-population based DE could help avoiding premature convergence in the

25 26 Chapter 5. Conclusion and Future Work early stages of the search. If one of the populations gets stuck into a local op- timum, a second population could help the first population to move out of the stagnation. Hence, combination of different mutation strategies via cooperative sub-populations could be investigated in the future research. Moreover, as mu- tation strategy is a sensitive factor to the problems of interest, a study of all the existing mutation strategies has to be done to acquire new knowledge of which mutation strategy would be most suitable in a given situation. Bibliography Another interesting work in the future would be integration of parameter adaptation into a memetic DE algorithm (with local search) to reach more competent DE performance in solving complex and large scale optimization problems. [1] E. Wari and W. Zhu. A survey on metaheuristics for optimization in food manufacturing industry. Aplied Soft Computing, 46:328–343, 2016.

[2] G.A. Bakare, G. Krost, G.K. Venayagamoorthy, and U.O. Aliyu. Compar- ative application of differential evolution and particle swarm techniques to reactive power and voltage control. In International Conference on In- telligent Systems Applications to Power Systems, 2007. ISAP 2007. Toki Messe, Niigata, pages 1–6, 2007.

[3] W. Gong and Z. Cai. Parameter optimization of pemfc model with im- proved multi-strategy adaptive differential evolution. Engineering Appli- cations of Artificial Intelligence, 27:28–40, 2014.

[4] J.H. Van Sickel, K.Y. Lee, and J.S. Heo. Differential evolution and its applications to power plant control. In International Conference on In- telligent Systems Applications to Power Systems, 2007. ISAP 2007., Toki MEsse, Niigata, pages 1–6. IEEE, 2007.

[5] S. Russel and P. Norvig. Artificial intelligence: A moder approach. In New Yersey: Prentice Hall, pages 111–114, 2003.

[6] R. Helena, O.M. Lourenzo, and T. Stutzle. Iterated local search. Springer, 2003.

[7] R. Storn and K. Price. Differential evolution - a simple and efficient heuristic for global optimization over continuous spaces. Journal of Global Optimization, 11(4):341 – 359, 1997.

[8] S. Kirkpatrick, C.D. Gelantt, and M.P. Vecchi. Optimization by simulated annealing. Science, 220:671–680, 1983.

[9] G. Glover. Tabu search - part one. ORSA Journal Computing, 1:190–206, 1989.

27 Bibliography

[1] E. Wari and W. Zhu. A survey on metaheuristics for optimization in food manufacturing industry. Aplied Soft Computing, 46:328–343, 2016.

[2] G.A. Bakare, G. Krost, G.K. Venayagamoorthy, and U.O. Aliyu. Compar- ative application of differential evolution and particle swarm techniques to reactive power and voltage control. In International Conference on In- telligent Systems Applications to Power Systems, 2007. ISAP 2007. Toki Messe, Niigata, pages 1–6, 2007.

[3] W. Gong and Z. Cai. Parameter optimization of pemfc model with im- proved multi-strategy adaptive differential evolution. Engineering Appli- cations of Artificial Intelligence, 27:28–40, 2014.

[4] J.H. Van Sickel, K.Y. Lee, and J.S. Heo. Differential evolution and its applications to power plant control. In International Conference on In- telligent Systems Applications to Power Systems, 2007. ISAP 2007., Toki MEsse, Niigata, pages 1–6. IEEE, 2007.

[5] S. Russel and P. Norvig. Artificial intelligence: A moder approach. In New Yersey: Prentice Hall, pages 111–114, 2003.

[6] R. Helena, O.M. Lourenzo, and T. Stutzle. Iterated local search. Springer, 2003.

[7] R. Storn and K. Price. Differential evolution - a simple and efficient heuristic for global optimization over continuous spaces. Journal of Global Optimization, 11(4):341 – 359, 1997.

[8] S. Kirkpatrick, C.D. Gelantt, and M.P. Vecchi. Optimization by simulated annealing. Science, 220:671–680, 1983.

[9] G. Glover. Tabu search - part one. ORSA Journal Computing, 1:190–206, 1989.

27 28 Bibliography

[10] M. Avriel. Nolinear programming: Analysis and methods. In Dover Publishing, 2003.

[11] J. A. Nelder and R. A. Mead. A simplex for function minimization, com- puter journal. Computer Journal, 7:308–313, 1965.

[12] N. Mladenovic and P. Hansen. Variable neighborhood search. Computers and Operations Research, 24(11):1097–1100, 1997.

[13] T.A. Feo and M.G.C. Resende. Greedy randrandom adaptive search pro- cedures. Journal of Global Optimizaiton, 6:109–134, 1995.

[14] R. Ruiz and T. Stutzle. A simple and effective iterated for the permutation flowshop scheduling problem. European Journal of Operational Research, 177(3):2033–2049, 2007.

[15] M.D. Vose. The simple genetic algorithm: foundations and theory, vol- ume 12. MIT Press, 1999.

[16] J. Kenedy and R. C. Eberhart. Particle swarm optimization. In In Proc. IEEE Conference on Neural Networks, pages 1942–1948, 1995.

[17] N Xiong, D. Molina, M.. Leon, and F. Herrera. A walk into metaheuris- tics for engineering optimization: Principles, methods, and recent trends. International Journal of Computational Intelligence Systems, 8(4):606– 636, 2015.

[18] Gu Jirong and Gu Guojun. Differential evolution with a local search op- erator. In Proc. 2010 2nd International Asia Conference on Informatics in Control, Automation and Robotics (CAR), Wuhan, China, volume 2, pages 480 – 483, 2010.

[19] Z. Dai and A. Zhou. A diferential ecolution with an orthogonal lo- cal search. In Proc. 2013 IEEE congress on Evolutionary Computation (CEC), Cancun, Mexico, pages 2329 – 2336, 2013.

[20] D. Molina, M. Lozano, C. Garcia-Martinez, and F. Herrera. Memetic al- gorithms for continuous optimization based on local search chains. Evo- lutionary Computation, 18:27–63, 2010.

[21] E. Ozcan and C. Basaran. A case study of meemtic algorithms for con- straint optimization. Soft Computing, 13:871–882, 2009. Bibliography 29

[22] B. Mohanty, S. Panda, and Hota PK. Controller parameters tuning of dif- ferential evolution algorithm and its application to load frequency control of multi-source power system. International journal of electrical power & energy systems, 54:77–85, 2014.

[23] E. Harth and E. Tzanakou. Alopex: A stochastic method for determining visual receptive fields. Vision Research, 14:1475–1482, 1974.

[24] E. Mezura-Montes, J. Velazquez-Reyes, and C. A. Coello Coello. A com- parative study of different evolution variants for global optimization. In Proc. Genetic and Evolutionary Computing Conference, pages 485–492, 2006.

[25] D. Zaharie. Influence of crossover on the behavior of differential evolution algorithms. Applied Soft Computing, 9:1126–1138, 2009.

[26] A.K. Qin and P.N. Suganthan. Self-adaptive differential evolution algo- rithm for numerical optimization. The 2005 IEEE Congress on Evolution- ary Computation, 2:1785–1791, 2005.

[27] Q. fan and X. Yan. Self-adaptive differential evolution algorithm with discrete mutation control parameters. Expert Systems with Applications, 42:1551–1572, 2015.

[28] X. Zhao, X. Wang, H. Sun, L. Wang, and M. Ma. A self-adaptive differ- ential evolution algorithm with an external archive for unconstrained op- timization problems. Journal of Intelligent and Fuzzy systems, 29:2193– 2204, 2015.

[29] J. Zhang and A.C. Sanderson. Jade: Adaptive differential evolution with optional external archive. IEEE Transactions on Evolutionary Computa- tion, 13:945–958, 2009.

[30] S. M. Islam, S. Das, S. Ghoshand, S. Roy, and P. N. Suganthan. An adap- tive differential evolution algorithm with novel mutation and crossover strategies for global numerical optimization. Systems, Man, and Cyber- netics, Part B: Cybernetics, IEEE Transactions on, 42(2):482–500, 2012.

[31] C. Wang, X. Wang, J. Xiao, and Y.Ding. Improved differential ecolution algorith based on dynamic adaptive strategies and control parameters. In- ternational Journal of Control and automation, 7(9):81–96, 2014. 30 Bibliography

[32] M. Yang, Z. Cai, C. Li, and J. Guan. An improved jade algorithm for global optimization. In IEEE Conference on Evolutionary Computation, pages 806–812, 2014.

[33] R. Tanabe and A. Fukinga. Success-history based parameter adaptation for differential evolution. In 2013 IEEE Congress on Evolutionary Com- putation (CEC), Cancun, Mexico, pages 71–78, 2013.

[34] R. Mallipeddi, P.N. P.N. Suganthan.and K.Q. Pan, and M.F. Tasgetiren. Differential evolution algorithm with ensemble of parameters and muta- tion strategies. Applied soft computing, 11(2):1679–1696, 2011.

[35] Y. Wang, Z. Cai, and Q. Zhang. Differential evolution with composite trial vector generation strategies and control parameters. IEEE Transaction on Evolutionary Computation, 15(1):55–66, 2011.

[36] J. Brest, S. Greiner, B. Boskovic, M. Mernik, and V. Zumer. Self-adapting control parameters in differential evolution: A comparative study on nu- merical benchmark problems. In IEEE Transaction on Evolutionary Com- putation, volume 10, pages 646–657, 2006.

[37] J. Liu and J. Lampinen. A fuzzy adaptive differential evolution algorithm. Soft Computing, 9(6):448–462, 2005.

[38] J. Teo. Exploring dynamic self-adaptive populations in differential evo- lution. Soft Comput.: Fusion Found., Methodologies Applicat.,, 10:673– 686, 2006.

[39] M. Lozano, F. Herrera, N. Krasnogor, and D. Molina. Real-coded memetic algorithms with crossover hill-climbing. Evolutionary Compu- tation, 12(3):273–302, 2004.

[40] H.G. Beyer and K. Deb. On self-adaptive ffeature in real-parameter evo- lutionary algorithms. IEEE Transaction on Evolutionary Computation, 5(3):250–270, 2001.

[41] B. Jiang. Optimizationion complex functions by chaos search. Cibernet- ics and Systems, 29(4), 1998.

[42] W. Pei-chong, Q. Xu, and H. Xiao-hong. A novel differential evolution algorithm based on chaos local search. In Proc. International conference on information Engineering and Computer Science, 2009. ICIECS 2009. Wuhan, China, pages 1–4, 2009. [43] D. Jia, G. Zheng, and M. K. Khan. An effective memetic differential evolution algorithm based on chaotical search. Information Sciences, 181:3175–3187, 2011.

[44] I. Poikolainen and F. Neri. Differential evolution with concurrent fitness based local search. In Proc. 2013 IEEE Congress on Evolutionary Com- putation (CEC), Cancun, Mexico, pages 384–391, 2013.

[45] N. Noman and H. Iba. Accelerating differential evolution using an adap- tative local search. In IEEE Transactions on Evolutionary Computation, volume 12, pages 107 – 125, 2008.

[46] N. Noman and Norman. Iba. Enhancing differential evolution perfor- mance with local search for high dimensional function optimization. In Proceedings of the 2005 conference on Genetic and evolutionary compu- tation, GECCO’05, pages 967–974, 2005.

[47] J. Gu and G. Gu. Differential evolution with a local search operator. In 2nd International Asia Conference on Informatics in Control and Robotics, pages 480–483, 2010.

[48] M. Ali, M. Pant, and A. Nagar. Two local search strategies for differential evolution. In Proc. Bio-Inspired Computing: Theories and Applications (BIC-TA), 2010 IEEE Fifth International Conference, Changsha, China, pages 1429 – 1435, 2010.

[49] H. Peng and Z. Wu. Heterozygous differential evolution with taguchi local search. Soft Computing, 19:3273–3291, 2015.

[50] A.K. Qin, K. Tang, H. Pang, and S. Xia. Self-adaptive differential evo- lution with local search chains for real-parameter single-objective opti- mization. In IEEE Congress on Evolutionary Computation (CEC), pages 467–474, 2014.