Investigating Recurrent Neural Network Memory Structures Using Neuro-Evolution

Total Page:16

File Type:pdf, Size:1020Kb

Investigating Recurrent Neural Network Memory Structures Using Neuro-Evolution Investigating Recurrent Neural Network Memory Structures using Neuro-Evolution Travis Desell ([email protected]) Associate Professor Department of Software Engineering Co-Authors: AbdElRahman ElSaid (PhD GRA) Alex Ororbia (Assistant Professor, RIT Computer Science) Collaborators: Steven Benson, Shuchita Patwardhan, David Stadem (Microbeam Technologies, Inc.) James Higgins, Mark Dusenbury, Brandon Wild (University of North Dakota) Overview • What is Neuro-Evolution? • Distributed Neuro- Evolution • Background: • Recurrent Neural • Results Networks for Time Series Prediction • Future Work • Recurrent Memory Cells • Discussion • EXAMM: • NEAT Innovations • Edge and Node Mutations • Crossover The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Motivation The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Neuro-Evolution • Applying evolutionary strategies to artificial neural networks (ANNs): • EAs to train ANNs (weight selection) • EAs to design ANNs (what architecture is best?) • Hyperparameter optimization (what parameters do we use for our backpropagation algorithm) The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Neuro-Evolution for Recurrent Neural Networks • Most people use human-designed ANNs, selecting from a few architectures that have done well in the literature. • No guarantees these are most optimal (e.g., in size, predictive ability, generalizability, robustness, etc). • Recurrent edges can go back farther in time than the previous time step -- dramatically increases the search space for RNN architectures. • With so many memory cell structures and architectures to choose from, which are best? What cells and architectures perform best, does this change across data sets and why? The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Neuro-Evolution • Applying evolutionary strategies to artificial neural networks (ANNs): • EAs to train ANNs (weight selection) • EAs to design ANNs (what architecture is best?) • Hyperparameter optimization (what parameters do we use for our backpropagation algorithm) • Using NE to better understand and guide ML: what structures and architectures are evolutionarily selected and why? The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Background The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Recurrent Neural Networks Hidden Hidden Input Hidden Hidden Output Input Output Node 1 Node 2 time 0 Node 1 Node 2 time 0 Recurrent Neural Networks can be extremely challenging to train due to the exploding/ Input Hidden Hidden Output vanishing gradients problem. In short, when time 1 Node 1 Node 2 time 1 training a RNN over a time series (via backpropagation through time), it needs to be completely unrolled over the time series. Input Hidden Hidden Output For the simple example above (blue arrows time 2 Node 1 Node 2 time 2 are forward connections, red are recurrent), backpropagating the error from time 3 reaches all the way back to input at time 0 (right). Even with this extremely simple RNN, Input Hidden Hidden Output we end up having an extremely deep time 3 Node 1 Node 2 time 3 network to train. ... The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Recurrent Neural Networks Hidden Hidden Input Hidden Hidden Output Input Output Node 1 Node 2 time 0 Node 1 Node 2 time 0 Traditionally, RNN connections go back a single time step and to the Input Hidden Hidden Output same or a previous node in the time 1 Node 1 Node 2 time 1 network. This is not a requirement - they can Input Hidden Hidden Output go back multiple time steps (green) time 2 Node 1 Node 2 time 2 or forward in the network (orange). However it is not as well studied due to additional complexity and Input Hidden Hidden Output dramatically increasing the time 3 Node 1 Node 2 time 3 architectural search space. ... The 2019 Genetic and Evolutionary Computation Conference Prague, Czech Republic July 13th-17th, 2019 Using ACO to Optimize LSTM Recurrent Neural Networks GECCO ’18, July 15–19, 2018, Kyoto, Japan t+10 Classification vs. Time Series Data Prediction Current input Table 1: K-Fold Cross Validation Results Input lag Actual vibration Prediction Errors (MAE) Hidden layer RNNs are perhaps more commonly used for classification (and have been mixed with CNNsLSTM for image NOE identification). NARX NBJ This ACO involves outputs being fed Subsamplethrough 1a softmax 8.34% 10.6%layer which 8.13% results 8.40% in 7.80%probabilities for the Subsample 2 4.05% 6.96% 6.08% 7.34% 3.70% input beingSubsample a particular 3 6.76% class. 16.8% The error 11.2% minimized 13.6% 3.49%is for the output being an incorrect class: t-9 t-1 t Mean 0.0638 0.1145 0.0847 0.0977 0.0501 Std. Dev. 0.0217 0.0497 0.0258 0.0333 0.0245 Figure 4: Nonlinear Output Error inputs neural network. 5.3 Error Function This network was updated to utilize 10 seconds of input data. For all the networks studied in this work, Mean Squared Error (MSE) t t+1 t+9 t+10 Current input RNNs can(shown also in be Equation used 1) for was time used asseries an error data measure prediction, for training, however in this Input lag case theas RNN it provides is predicting a smoother optimization an exact surface value for of backpropagation a time series, some Prediction than mean average error. Mean Absolute Error (MAE) (shown in Hidden layer number of time steps in the future. The error being minimized is Equation 2) was used as a nal measure of accuracy for the three typically architectures,the mean assquared because the error parameters (1) or were mean normalized absolute between error (2). This is an important0 and 1,distinction. the MAE is also the percentage error. 0.5 Actual V ib Predicted Vib 2 t-9 t-1 t Error = ⇥ ( − ) (1) Testin Seconds Õ ABS Actual V ib Predicted Vib Figure 5: Nonlinear AutoRegressive with eXogenous inputs Error = [ ( − )] (2) Testin Seconds neural network. This network was updated to utilize 10 sec- Õ onds of input data, along with the previous 10 predicted out- 5.4 Machine Specications The 2019 Genetic and Evolutionary Computation Conference put values. Prague, Czech Republic Python’s Theano Library [23] was used to implement the neural July 13th-17th, 2019 t t+1 t+9 t+10 networks and MPI for Python [3] and was used to run the ACO op- Actual vibration timization on a high performance computing cluster. The cluster’s Error Current input operating system was Red Hat Enterprise Linux (RHEL) 7.2, and Input lag Prediction had 31 nodes, each with 8 cores (248 cores in total) and 64GBs RAM Hidden layer (1948 GB in total). The interconnect was 10 gigabit (GB) InniBand. 6 RESULTS The ACO algorithm was run for 1000 iterations using 200 ants. The t-9 t-1 t networks were allowed to train for 575 epochs to learn and for the error curve to atten. The minimum value for the pheromones Figure 6: Nonlinear Box-Jenkins inputs neural network. were 1 and the maximum was 20. The population size was equal to This network was updated to utilize 10 seconds of input data, number number of iterations in the ACO process, i.e., the population along with the future output and error values. Due to requir- size was also 1000. Each run took approximately 4 days. ing future knowledge, it is not possible to utilize this net- A dataset of 57 ights was divided into 3 subsamples, each con- work in an online fashion. sisting of 19 ights. The subsamples were used to cross validate the results by examining combations utilizing two of the subsamples indicative of a case of vanishing gradients. Accordingly, the study as the training data set and the third as the testing set. Subsamples allowed for the recurrent weights to be considered in the gradient 1, 2 and 3 consisted of 23,371, 31,207 and 25,011 seconds of ight calculations in order to update the weights with respect to the cost data, respectively. function output. These subsamples were used to train the NOE, NARX, NBH, base 5.2.3 Nonlinear Box-Jenkins (NBJ) Inputs Neural Network: The architecture and the ACO optimized architecture. Figures 7 shows structure of the NBJ is depicted in Figure 6. As previously noted, predictions for the dierent models over a selection of test ights, this network is not feasible for prediction past one time step in and Figure 8 shows predictions an single uncompressed (higher the future in an online manner, as it requires the actual prediction resolution) test ight. Table 1 compares these models to the base value and error between it and the predicted value to be fed back architecture (LSTM) and the ACO optimized architecture (ACO). into the network. However, as this work delt with oine data, the actual future vibration values, error, and the output were all fed to 6.1 NOE, NARX, and NBJ Results the network along with the current instance parameters and lag Somewhat expectedly, the NOE model performed the worst with inputs. As in the other networks, the values for the previous 10 with a mean error of 11.45% (σ = 0.0497). The NBJ model performed time steps were also utilized. better than the NOE model with a mean error of 9.77% (σ = 0.0333), Memory Cell Structures - Simple Neuron • "Simple" neurons are the basic neural network building block. Inputs are summed, and activation Input function is applied and the (sum) output is feed forward to other neurons. • Simple (and any other Activation Function memory cell structure) can (tanh, also have recurrent edges (in sigmoid, ...) red) added through the evolutionary process.
Recommended publications
  • Green Parallel Metaheuristics: Design, Implementation, and Evaluation
    DOCTORAL THESIS Green Parallel Metaheuristics: Design, Implementation, and Evaluation PhD Thesis Dissertation in Computer Science Author Amr Abdelhafez Directors Enrique Alba Torres and Gabriel Luque Polo E. T. S. I. Informática Lenguajes y Ciencias de la Computación UNIVERSIDAD DE MÁLAGA (España) April 2020 AUTOR: Amr Abdelhafez Abdelsamee Galal http://orcid.org/0000-0003-0469-3791 EDITA: Publicaciones y Divulgación Científica. Universidad de Málaga Esta obra está bajo una licencia de Creative Commons Reconocimiento-NoComercial- SinObraDerivada 4.0 Internacional: http://creativecommons.org/licenses/by-nc-nd/4.0/legalcode Cualquier parte de esta obra se puede reproducir sin autorización pero con el reconocimiento y atribución de los autores. No se puede hacer uso comercial de la obra y no se puede alterar, transformar o hacer obras derivadas. Esta Tesis Doctoral está depositada en el Repositorio Institucional de la Universidad de Málaga (RIUMA): riuma.uma.es DECLARACIÓN DE AUTORÍA Y ORIGINALIDAD DE LA TESIS DOCTORAL Considerando que la presentación de un trabajo hecho por otra persona o la copia de textos, fotos y gráficas sin citar su procedencia se considera plagio, el abajo firmante Don Amr Abdelhafez con NIE: X que presenta la tesis doctoral con el título: Metaheurísticas Paralelas Verdes: Diseño, Implementación y Evaluación, declara la autoría y asume la originalidad de este trabajo, donde se han utilizado distintas fuentes que han sido todas citadas debidamente en la memoria. Y para que así conste firmo el presente documento en Málaga a……………….. El autor: ………………….….. Departamento de Lenguajes Y Ciencias de La Computación Escuela Técnica Superior de Ingeniería Informática Universidad de Málaga El Dr.
    [Show full text]
  • New Approaches in Parallel Algorithm Portfolios for Metaheuristic Optimization
    New Approaches in Parallel Algorithm Portfolios for Metaheuristic Optimization A Dissertation submitted to the designated by the General Assembly of Special Composition of the Department of Computer Science and Engineering Examination Committee by Dimitrios Souravlias in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY University of Ioannina June 2017 Advisory Committee • Konstantinos E. Parsopoulos, Associate Professor, Department of Computer Science & Engineering, University of Ioannina, Greece • Enrique Alba, Professor, Department of Computer Science & Programming Languages, University of Málaga, Spain • Konstantina Skouri, Associate Professor, Department of Mathematics, Univer- sity of Ioannina, Greece Examination Committee • Konstantinos E. Parsopoulos, Associate Professor, Department of Computer Science & Engineering, University of Ioannina, Greece • Enrique Alba, Professor, Department of Computer Science & Programming Languages, University of Málaga, Spain • Konstantina Skouri, Associate Professor, Department of Mathematics, Univer- sity of Ioannina, Greece • Isaac Lagaris, Professor, Department of Computer Science & Engineering, Uni- versity of Ioannina, Greece • Ilias Kotsireas, Professor, Department of Physics & Computer Science, Wilfrid Laurier University, Canada • Vassilios Dimakopoulos, Associate Professor, Department of Computer Science & Engineering, University of Ioannina, Greece • Dimitrios Papageorgiou, Associate Professor, Department of Materials Science & Engineering, University of Ioannina, Greece Dedication To my family Acknowledgements First and foremost I would like to express my deepest gratitude to my advisor, Professor Konstantinos E. Parsopoulos, for his valuable guidance and continuous support from the beginning of my PhD studies. Our excellent collaboration has given me the opportunity to acquire important knowledge and skills that have helped me in my research endeavor. Our long discussions and his constant encouragement have greatly contributed to my scientific and personal development.
    [Show full text]
  • Deep Neuroevolution: Smart City Applications
    UNIVERSIDAD DE MÁLAGA DOCTORAL THESIS Deep Neuroevolution: Smart City Applications Author: Supervisor: M.Sc. Andrés CAMERO Prof. Dr. Enrique ALBA A thesis submitted in fulfillment of the requirements for the degree of Doctor of Philosophy in the Programa de Doctorado en Tecnologías Informáticas Departmento de Lenguajes y Ciencias de la Compuación E.T.S.I. Informática February 12, 2021 ! AUTOR: Andrés Camero Unzueta http://orcid.org/0000-0002-8152-9381 EDITA: Publicaciones y Divulgación Científica. Universidad de Málaga ! Esta obra está bajo una licencia de Creative Commons Reconocimiento-NoComercial- SinObraDerivada 4.0 Internacional: http://creativecommons.org/licenses/by-nc-nd/4.0/legalcode Cualquier parte de esta obra se puede reproducir sin autorización pero con el reconocimiento y atribución de los autores. No se puede hacer uso comercial de la obra y no se puede alterar, transformar o hacer obras derivadas. Esta Tesis Doctoral está depositada en el Repositorio Institucional de la Universidad de Málaga (RIUMA): riuma.uma.es iii M.Sc. Andrés Camero Unzueta, estudiante del programa de doctorado Tec- nologías Informáticas de la Universidad de Málaga, autor de la tesis, presentada para la obtención del título de doctor por la Universidad de Málaga, titulada: Deep Neuroevolution: Smart City Applications Realizada bajo la tutorización y dirección del Prof. Dr. Enrique Alba Torres. Declaro que La tesis presentada es una obra original que no infringe los derechos de propiedad intelectual ni los derechos de propiedad industrial u otros, conforme al ordenamiento jurídico vigente (Real Decreto Legislativo 1/1996, de 12 de abril, por el que se aprueba el texto refundido de la Ley de Propiedad Intelectual, regularizando, aclarando y armonizando las disposiciones legales vigentes sobre la materia), modificado por la Ley 2/2019, de 1 de marzo.
    [Show full text]
  • Systolic Genetic Search, a Parallel Metaheuristic for Gpus
    Tesis de Doctorado en Informática Systolic Genetic Search, A Parallel Metaheuristic for GPUs Martín Pedemonte Director de Tesis: Prof. Enrique Alba Torres Co-Director de Tesis: Prof. Francisco Luna Valero Director Académico: Prof. Héctor Cancela PEDECIBA Informática Instituto de Computación – Facultad de Ingeniería Universidad de la República Setiembre 2017 “Life is what happens to you, while you’re busy making other plans” John Lennon, Beautiful Boy (Darling Boy) To Juan Pablo and Isabel, two by-products of this thesis. Agradecimientos Finalmente ha llegado el día y el Doctorado está terminando. Este momento nunca hubiera sido posible sin la colaboración de un conjunto de personas, que me apoyaron a lo largo de este proceso, a las que quiero expresar mi agradecimiento. En primer lugar, quiero agradecer a mis directores de Tesis Enrique y Paco por su guía académica, su aporte a este trabajo, y la confianza en mí y en mi trabajo. También agradezco a Héctor, quien fue mi director académico, por su ayuda y su estímulo. A Pablo y Ernesto por aportarme otras perspectivas, por los intercambios de ideas y por sus valiosos comentarios que contribuyeron con este trabajo. A Eduardo por sus sugerencias e ideas para el diseño de una nueva grilla. Agradezco a mi compañera Leonella por su apoyo incondicional y su paciencia infinita, y también por haberme facilitado las cosas para que pudiera dedicarle tiempo a esta tesis. También agradezco a mis hijos Juan Pablo e Isabel por todos los ratos que les tuve que robar. A mi madre Graciela y a mi hermano Gerardo por su continuo aliento y por su cariño.
    [Show full text]
  • CURRICULUM VITAE Carlos Artemio Coello Coello February 27Th, 2021
    CURRICULUM VITAE Carlos Artemio Coello Coello February 27th, 2021 Personal Information DATE OF BIRTH : October 18, 1967 PLACE OF BIRTH : Tonala,´ Chiapas, Mexico´ PHONE : +52 (55) 5747 3800 x 6564 EMAIL : [email protected] WORLD WIDE WEB : http://delta.cs.cinvestav.mx/˜ccoello CURRENT WORKPLACE: Departamento de Computacion´ CINVESTAV-IPN, Av. IPN No. 2508, Mexico,´ D.F. 07360, Mexico´ CURRENT POSITION: Investigador CINVESTAV 3F (Professor with Distinction) ORCID ID: 0000-0002-8435-680X ResearcherID: B-1845-2008 Scopus Author ID: 7003514400 Academic Degrees • BSc in Civil Engineering, Universidad Autonoma´ de Chiapas, 1985–1990. Degree awarded in 1991 with the thesis “Analysis of Grid Structures using a personal computer (stiffness method)” (summa cum laude). • Master of Science in Computer Science, Tulane University, New Orleans, Louisiana, EUA, 1991–1993. Degree awarded in December, 1993. • PhD in Computer Science, Tulane University, New Orleans, Louisiana, EUA, 1993–1996. Degree awarded in May, 1996, with the thesis entitled “An Empirical Study of Evolutionary Techniques for Multiobjective Optimization in Engineering Design”. Advisor: Dr. Alan D. Christiansen. Most Relevant Publications Monographic Books 1. Carlos A. Coello Coello, David A. Van Veldhuizen and Gary B. Lamont, Evolutionary Algorithms for Solving Multi-Objective Problems, Kluwer Academic Publishers, New York, USA, May 2002, ISBN 0-3064-6762-3. 2. Carlos A. Coello Coello, Gary B. Lamont and David A. Van Veldhuizen, Evolutionary Algorithms for Solving Multi-Objective Problems, Springer, New York, USA, Second Edition, ISBN 978-0-387-33254-3, September 2007. Edited Books 1. Carlos A. Coello Coello and Gary B. Lamont (eds), Applications of Multi-Objective Evolutionary Algorithms, World Scientific, Singapore, 2004, ISBN 981-256-106-4.
    [Show full text]
  • In This Issue
    Volume 7 SIGEVOLution Issue 4 newsletter of the ACM Special Interest Group on Genetic and Evolutionary Computation in this issue GECCO 2015 Summary GECCO best paper winners Latest contents from ECJ and GPEM CFPs for upcoming conferences Latest vacancies Track: Estimation of Distribution Algorithms (EDA)+ EDITORIAL THEORY Those of you who managed to make it to sunny Winner: Improved Runtime Bounds for the (1+1) Madrid for GECCO this year will already know EA on Random 3-CNF Formulas Based on Fitness- what a fantastic conference it was from every Distance Correlation perspective – great science, great company and Benjamin Doerr, Frank Neumann, Andrew M. great food! For those who missed it, you can catch Sutton (Full text) up with best paper awards and read a summary from this year’s General Chair Anna Esparcia- Track: Evolutionary Combinatorial Optimization Alcazar inside. You will also find the latest tables and Metaheuristics (ECOM) of contents from ECJ and GPEM, as well as calls Winner: Global vs Local Search on Multi-objective for some of the conferences coming up in 2016. NK-landscapes: Contrasting the Impact of Problem Get writing those papers! The next issue of the Features newsletter will provide an insider’s view on Denver Fabio Daolio, Arnaud Liefooghe, Sébastien Verel, which will be the location for GECCO 2016, just in Hernán Aguirre, Kiyoshi Tanaka (Full text) case you need any persuasion to attend. I’d also like to take this opportunity to congratulate Marc Track: Evolutionary Machine Learning -EML Schoenauer on becoming Chair of SIGEVO from July and to extend a warm thank you to Wolfgang Winner: Subspace clustering using evolvable Banzhaf for his hard work over the last four years.
    [Show full text]
  • Bayesian Neural Architecture Search Using a Training-Free Performance Metric
    Bayesian Neural Architecture Search using A Training-Free Performance Metric Andr´esCamerob,a,∗, Hao Wangc, Enrique Albaa, Thomas B¨ackc aUniversidad de M´alaga,ITIS Software, Espa~na bGermanAerospace Center (DLR), Remote Sensing Technology Institute (IMF), Germany cLeiden University, LIACS, The Netherlands Abstract Recurrent neural networks (RNNs) are a powerful approach for time se- ries prediction. However, their performance is strongly affected by their architecture and hyperparameter settings. The architecture optimization of RNNs is a time-consuming task, where the search space is typically a mixture of real, integer and categorical values. To allow for shrinking and expanding the size of the network, the representation of architectures often has a variable length. In this paper, we propose to tackle the architecture optimization problem with a variant of the Bayesian Optimization (BO) algorithm. To reduce the evaluation time of candidate architectures the Mean Absolute Error Random Sampling (MRS), a training-free method to estimate the network performance, is adopted as the objective function for BO. Also, we propose three fixed-length encoding schemes to cope with the variable-length architecture representation. The result is a new perspective on accurate and efficient design of RNNs, that we validate on three problems. Our findings show that 1) the BO algorithm can explore different network architectures using the proposed encoding schemes and successfully designs well-performing architectures, and 2) the optimization time is significantly
    [Show full text]
  • Candidate for Executive Committee Enrique Alba Universidad De
    Candidate for Executive Committee Enrique Alba Universidad de Malaga, Spain BIOGRAPHY Academic Background: Ph.D. Computer Science, University of Malaga (Spain), 1999, Parallel Advanced GAs for Neural Network Design. Professional Experience: Full Professor, Universidad de Malaga, Spain, 2009 – Present; Associate Professor, Universidad de Malaga, Spain, 2001 – 2009; Assistant Professor, Universidad de Malaga, Spain, 1993 – 2001. Professional Interest: Intelligent Systems for Smart Cities; High Performance Computing for Artificial Intelligence; Software Engineering and Artificial Intelligence; Advanced Algorithms: Multiobjective, Hybrid, Dynamic Machine Learning, Search & Complex Optimization in Real Problems. ACM Activities: Executive Board Member, SIGEVO, 2014 – Present. Membership and Offices in Related Organizations: Executive Commission, Spanish AI organization for National Master on AI, 2016 – 2018; ERC reviewer, Horizon 2020 evaluator, European Union, 2000 – Present; CLAIRE initiative for AI in Europe, European Union, 2018 – Present. Awards Received: Best papers at ACM GECCO MO Track and in INCoS 2016; Best Paper in Conference EVOCOP, 2007; Best Paper in Conference PPSN, 2006; National Award "Jose Garcia Santesmases" best Graduate Thesis, 1993. STATEMENT I have been dealing with complex problems and different types of AI during all my 25 years of research. I hold links and active collaborations in all continents, working hard to promote Evolutionary Computing and related fields at graduate and postgraduate levels. My intention is to help ACM SIGEVO by using my experience and network to endorse the main goals of this SIG throughout the world, both in research and in teaching. I would also like to stress the importance of transferring knowledge from academia to industry, to have a positive impact in the lives of people.
    [Show full text]
  • Conference Program
    Genn etic and Evolutionary Computation Conference (GECCO-2004) CONFERENCE PROGRAM June 26 – 30, 2004 Red Lion Hotel on Fifth Avenue Seattle, Washington, USA A Recombination of the Ninth Annual Genetic Programming Conference and the Thirteenth International Conference on Genetic Algorithms International Society for Genetic and Evolutionary Computation, Inc. In Association with the American Association for Artificial Intelligence Table of Contents GECCO Organizers ................................................................................................................................................... 2 Program Committee................................................................................................................................................... 3 Best Paper Awards..................................................................................................................................................... 4 About the Evolutionary Computation in Industry Track ........................................................................................... 5 Hotel Floor Plan......................................................................................................................................................... 5 Saturday Workshops and Tutorials............................................................................................................................ 6 Sunday Workshops, Tutorials and Human-competitive Results ..............................................................................
    [Show full text]
  • Hadoop, Spark and MPI
    Big Optimization with Genetic Algorithms: Hadoop, Spark and MPI Carolina Salto Universidad Nacional de La Pampa Gabriela Minetti ( [email protected] ) Universidad Nacional de La Pampa https://orcid.org/0000-0003-1076-6766 Enrique Alba Universidad de Málaga Gabriel Luque Universidad de Málaga: Universidad de Malaga Research Article Keywords: Big optimization, Genetic Algorithms, MapReduce, Hadoop, Spark, MPI Posted Date: July 31st, 2021 DOI: https://doi.org/10.21203/rs.3.rs-725766/v1 License: This work is licensed under a Creative Commons Attribution 4.0 International License. Read Full License Noname manuscript No. (will be inserted by the editor) 1 2 3 4 5 6 Big Optimization with Genetic Algorithms: Hadoop, Spark 7 8 and MPI 9 10 Carolina Salto · Gabriela Minetti · Enrique Alba · Gabriel Luque 11 12 13 14 15 16 17 18 the date of receipt and acceptance should be inserted later 19 20 21 Abstract Solving problems of high dimensionality (and benefits of these systems, namely file, memory and com- 22 complexity) usually needs the intense use of technolo- munication management, over the resulting algorithms. 23 gies, like parallelism, advanced computers and new types We analyze our MRGA solvers from relevant points of 24 of algorithms. MapReduce (MR) is a computing paradigm view like scalability, speedup, and communication vs. 25 long time existing in computer science that has been computation time in big optimization. The results for 26 proposed in the last years for dealing with big data ap- high dimensional datasets show that the MRGA over 27 plications, though it could also be used for many other Hadoop outperforms the implementations in Spark and 28 tasks.
    [Show full text]
  • Table of Contents
    Table of Contents EvoCOMNET Contributions Web Application Security through Gene Expression Programming 1 Jaroslaw Skaruz and Franciszek Seredynski Location Discovery in Wireless Sensor Networks Using a Two-Stage Simulated Annealing 11 Guillermo Molina and Enrique Alba Wireless Communications for Distributed Navigation in Robot Swarms 21 Gianni A. Di Caro, Frederick Ducatelle, and Luca M. Gambardella An Evolutionary Algorithm for Survivable Virtual Topology Mapping in Optical WDM Networks 31 Fatma Corut Ergin, Aygegiil Yayimli, and §ima Uyar Extremal Optimization as a Viable Means for Mapping in Grids 41 Ivanoe De Falco, Antonio Delia Cioppa, Domenico Maisto. Umberto Scafuri, and Ernesto Tarantino Swarm Intelligence Inspired Multicast Routing: An Ant Colony Optimization Approach 51 Xiao-Min Hu, Jun Zhang, and Li-Ming Zhang A Framework for Evolutionary Peer-to-Peer Overlay Schemes 61 Michele Amoretti Multiuser Scheduling in HSDPA with Particle Swarm Optimization .... 71 Mehmet E. Aydin, Raymond Kwan, Cyril Leung, and Jie Zhang Efficient Signal Processing and Anomaly Detection in Wireless Sensor Networks 81 Markus Wdlchli and Torsten Braun Peer-to-Peer Optimization in Large Unreliable Networks with Branch-and-Bound and Particle Swarms 87 Baldzs Bdnhelyi, Marco Biazzini, Alberto Montresor. and Mark Jelasity Evolving High-Speed, Easy-to-Understand Network Intrusion Detection Rules with Genetic Programming 93 Agustin Orfila, Juan M. Estevez-Tapiador. and Ariuro Ribagorda Bibliografische Informationen digitalisiert durch http://d-nb.info/993136311 XXIV Table of Contents Soft Computing Techniques for Internet Backbone Traffic Anomaly Detection 99 Antonia Azzini, Matteo De Felice, Sandro Meloni, and Andrea G.B. Tettamanzi Testing Detector Parameterization Using Evolutionary Exploit Generation 105 Hilmi G.
    [Show full text]
  • Enrique Alba Referencias a Mis Trabajos (References to My Papers) Feb 7, 2007 Y Parallelism and Evolutionary Algorithms – E
    Enrique Alba Referencias a mis Trabajos (References to my Papers) Feb 7, 2007 y Parallelism and Evolutionary Algorithms – E. Alba and M. Tomassini, IEEE Transactions on Evolutionary Computation 6(5):443-462, 2002 1. A Scalable Parallel Genetic Algorithm for X-ray Spectroscopic Analysis, K. Xu, S.J. Louis, R.C. Manzini, ACM Press, en Procs. of GECCO’05, pp. 811-816, 2005 2. Un Algoritmo Evolutivo Multiobjetivo Paralelo Aplicado al Diseño de Redes de Comunicaciones Confiables, S. Nesmachnow, en Actas de MAEB’05, Thomson, pp. 145-152, 2005 3. A Takeover Time Curves in Random and Small-World Structured Populations, M. Giacobini, M. Tomassini, A. Tettamanzi, ACM Press, en Procs. of GECCO’05, pp. 1333-1340, 2005 4. Parallel Genetic Algorithm for Graph Coloring Problem, Z. Kokosinski, M. Kolodziej, K. Kwarciany 5. A Parallel Approach to Elevator Optimization Based on Soft Computing, T. Beielstein, S. Markon, and M. Preu. In T. Ibaraki (ed.) Procs. 5th Metaheuristics Int'l Conf. (MIC'03), pp. 07/1--07/11 (CD- ROM), Kyoto, Japan, 2003 6. Considerations in Engineering Parallel Multiobjective Evolutionary Algorithms, Van Veldhuizen DA, Zydallis JB, Lamont GB, IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION 7 (2): 144-173 APR 2003 7. A Specialized Island Model and its Application in Multiobjective Optimization, Xiao NC, Armstrong MP, LECTURE NOTES IN COMPUTER SCIENCE 2724: 1530-1540 2003 8. Evolutionary Design of Rule Changing Cellular Automata, Kanoh H, Wu Y, LECTURE NOTES IN ARTIFICIAL INTELLIGENCE 2773: 258-264 2003 9. Ten Years of Genetic Fuzzy Systems: Current Framework and New Trends, Cordon O, Gomide F, Herrera F, et al.
    [Show full text]