Assessment of the Modeling Abilities of Neural Networks Can Be Formed

Total Page:16

File Type:pdf, Size:1020Kb

Assessment of the Modeling Abilities of Neural Networks Can Be Formed ASSESSMENT OF THE MODELING ABILITIES OF NEURAL NETWORKS by Alvin Ramsey Submitted to the Department of Mechanical Engineering in Partial Fulfillment of the Requirements for the Degree of Master of Science in Mechanical Engineering at the Massachusetts Institute of Technology January 1994 © Massachusetts Institute of Technology 1994 All rights reserved Signature of Author .. lDpartment of Mechanical Engineering January 1994 /t / I I Certified by --- George Chryssolouris -~~ 1 I ,'4ro . essor Thesis Supervisor l l V~rto Accepted by - ............- ... -'-- Professor Ain A. Sonin i A)-:S1'!!3NSr Graduate Committee OFTFrHR A 5,,'RARIE DEDICATION To my mother ( o o3 ) and to my brother ( Bro, on my word, we shall go fishing someday! ). 2 ASSESSMENT OF THE MODELING ABILITIES OF NEURAL NETWORKS by Alvin Ramsey Submitted to the Department of Mechanical Engineering on January 1994 in partial fulfillment of the requirements for the Degree of Master of Science in Mechanical Engineering ABSTRACT The treatment of manufacturing problems, whether in process control, process optimization, or system design and planning, can be helped by input-output models, namely, relationships between input and output variables. Artificial neural networks present an opportunity to "learn" empirically established relationships and apply them subsequently in order to solve a particular problem. In light of the increasing amount of applications of neural networks, the objective of this thesis is to evaluate the ability of neural networks to generate accurate models for manufacturing applications. Various neural network models has been tested on a number of "test bed" problems which represent the problems typically encountered in manufacturing processes and systems to assess the reliability of neural network models and to determine the efficacy of their modeling capabilities. The first type of problem tested on neural networks is the presence of noise in experimental data. A method to estimate the confidence intervals of neural network models has been developed to assess their reliability, and the proposed method has succeeded for a number of the models of the test problems in estimating the reliability of the neural network models, and greater accuracy may be achieved with higher-order calculations of confidence intervals which would entail increased computational burden and a higher requirement of precision for the parametric values of the neural network model. The second type of problem tested on neural networks is the high level of nonlinearity typically present in an input-output relationship due to the complex phenomena associated within the process or system. The relative efficacy of neural net modeling is evaluated by comparing results from the neural network models of the test bed problems with results from models generated by other common modeling methods: linear regression, the Group Method of Data Handling (GMDH), and the Multivariate Adaptive Regression Splines (MARS) method. The relative efficacy of neural networks has been concluded to be relatively equal to the empirical modeling methods of GMDH and MARS, but all these modeling methods are likely to give a more accurate model than linear regression. Thesis Supervisor: Professor George Chryssolouris Title: Associate Professor of Mechanical Engineering 3 ACKNOWLEDGMENTS Many deserves thanks and recognition for helping me develop the work performed for the Laboratory for Manufacturing and Productivity at MIT. I would like to thank my advisor, Professor George Chryssolouris, for his guidance, support, and opportunity to work in his research group. I must also thank Dr. Moshin Lee, Milan Nedeljkovic, Chukwuemeka (Chux) Amobi, and German Soto for their contributions to the work presented in this thesis; they are among the "buddies" of the lab. I like to thank the rest of the buddies: Subramaniam (the original Buddy), Andrew Yablon, Andrew Heitner, Sean Arnold, Eric Hopkins, Wolfram Weber, Johannes Weis, Giovanni Perrone, Dirk Loeberman, Ulrich Hermann, Helen Azrin, Tina Chen, and Steve Marnock. And I finally would like to thank my mother, Kun Cha Ramsey, and my brother, Dan Ramsey, for their continuous support. 4 TABLE OF CONTENTS LIST OF FIGURES .................................................................................. 7 LIST OF TABLES .................................................................................... 9 1. INTRODUCTION ............................................................................... 10 2. NEURAL NETWORKS FOR EMPIRICAL MODELING ................................ 15 2.1. Empirical vs. Analytical Modeling ...................................................... 15 2.2. Neural Network Background ........................................................ 16 2.2.1. The Pursuit of Pattern Recognition .............................................. 17 2.2.2. Backpropagation Neural Networks .............................................. 19 2.2.3 Other Neural Network Architectures ............................................. 22 3. NEURAL NETWORKS IN MANUFACTURING . .......................... 24 3.1. Areas of Applicability ........................................................ 24 3.1.1. Process Control, Diagnostics and Optimization ............................ 24 3.1.2. System Design and Planning .................. .................... 28 3.2. Current Applications of Neural Networks in Manufacturing ........................ 30 4. CONFIDENCE INTERVAL PREDICTION FOR NEURAL NETWORKS........................................................ 37 4.1. Confidence Intervals for Parameterized Models ............. ......................... 37 4.1.1. Selection of the Variance-Covariance Matrix ................................... 40 4.2. Derivation of Confidence Intervals for Models Derived from Noisy Data ......... 41 4.3. Derivation of Confidence Intervals for Neural Networks ............................ 43 5. APPLICATION OF CONFIDENCE INTERVAL PREDICTION FOR NEURAL NETWORKS ............................................... ................. 47 5.1. Test Problem: Multivariate Test Functions . .............................. 47 5.1.2. Neural Network Modeling of the Multivariate Test Functions ............... 48 5.1.3. Neural Network Reliability ................... .................................... 52 5.2. Test Problem: Laser Machining ........................................................ 53 5.2.1.Background ......... .. ......... .. ......... ......................53 5 5.2.2. Neural Network Modeling of Laser Through-Cutting ........................ 56 5.2.2. Neural Network Modeling of Laser Grooving ................................ 57 5.2.3. Neural Network Reliability ......... ........................................... 59 5.3. Test Problem: The Manufacturing of an Automobile Steering Column ........... 59 5.3.1. An Evaluation Framework for Manufacturing Systems ....................... 60 5.3.2. The Manufacturing of an Automobile Steering Column ................ 62 5.3.3. Neural Network Modeling of the Manufacturing of the Steering Column ............................................................................. 70 5.3.4. Neural Network Reliability ............................................. ...... 71 6. RELATIVE PERFORMANCE OF NEURAL NETWORKS .............................. 73 6.1. Linear Regression ................................................................ 73 6.2. Group Method of Data Handling ....................................................... 74 6.3. Multivariate Adaptive Regression Splines ........... ....................... 78 6.3.1. Recursive Partitioning Regression . ............................. 79 6.3.2. Adaptive Regression Splines ........ ...................... 80 6.4. Results of the Various Modeling Approach to the Test Bed Problems ........... 81 6.4.1. Modeling of the Multivariate Test Functions ............................. 82 6.4.2. Modeling of Laser Machining .......... .... ..... .................................85 6.4.2.1. Laser Through-Cutting ..................................................... 85 6.4.2.2.Laser Grooving ........... ........ .. .....................86 6.4.3. Modeling of the Manufacturing of an Automobile Steering Column ........ 87 6.4.4. Overall Assessment of the Relative Modeling Efficacy of Neural Networks .......................................................................... 89 7. CONCLUSION ...................................... 900................... 7.1. Use of Confidence Intervals to Assess Neural Network Reliability ........ 90 7.2. Relative Efficacy of Neural Network Models ......................................... 91 7.3. Summary .................................................................. 92 REFERENCES ...................................................... ........................ 93 6 LIST OF FIGURES Figure 2.1: Example of a linearly separable pattern ................... 1......................18 Figure 2.2: The exclusive OR problem ................... 1....................................19 Figure 2.3: An example of a feedforward neural network .......... .................. 20 Figure 2.4: A 1-1-1neural network with linear input and output nodes ................... 22 Figure 4-1: A 1-1-1 neural network with linear input and output nodes ................... 43 Figure 5-1: 806%confidence intervals for the neural network model of function 1........................................................... 49 Figure 5-2: 800%confidence intervals for the neural network model of function 1 containing noise ......................................................
Recommended publications
  • Neural Networks, Biometrics and the New Artificial Intelligence Mathias Fuchs & Ramón Reichert 5
    ISBN: 2838 Digital Vol. 4, Issue 1/2018 Ramón Reichert, Mathias Fuchs, Pablo Abend, Annika Richterich, Karin Wenz (eds.) Rethinking AI: Neural Networks, Biometrics and the New Artifical Intelligence Ramón Reichert, Annika Richterich (eds.) Digital Material/ism The journal is edited by Annika Richterich, Karin Wenz, Pablo Abend, Mathias Fuchs, Ramón Reichert Editorial Board Maria Bakardjieva, Brian Beaton, David Berry, Jean Burgess, Mark Coté, Colin Cremin, Sean Cubitt, Mark Deuze, José van Dijck, Delia Dumitrica, Astrid Ensslin, Sonia Fizek, Federica Frabetti, Richard A. Grusin, Orit Halpern, Irina Kaldrack, Wendy Hui Kyong Chun, Denisa Kera, Lev Manovich, Janet H. Murray, Jussi Parikka, Lisa Parks, Christiane Paul, Dominic Pettman, Rita Raley, Richard Rogers, Julian Rohrhuber, Marie-Laure Ryan, Mirko Tobias Schäfer, Jens Schröter, Trebor Scholz, Tamar Sharon, Roberto Simanowski, Nathaniel Tkacz, Nanna Verhoeff, Geoffrey Winthrop-Young, Sally Wyatt Bibliographic information published by the Deutsche Nationalbibliothek The Deutsche Nationalbibliothek lists this publication in the Deutsche National bibliografie; detailed bibliographic data are available on the Internet at http://dnb.d-nb.de © 2018 transcript Verlag, Bielefeld All rights reserved. No part of this book may be reprinted or reproduced or utilized in any form or by any electronic, mechanical, or other means, now known or hereafter invented, including photocopying and recording, or in any information storage or retrieval system, without permission in writing from the publisher.
    [Show full text]
  • Artificial Intelligence & the Future of Data-Driven Advertising
    Artificial Intelligence & The Future of Data-Driven Advertising Dr. Jochen Schlosser Chief Strategy Officer [email protected] 1 Dr. Jochen Schlosser Chief Strategy Officer Adform Owner of Adform’s strategy with dedicated responsibility for Product & Technology. • Broad Paid / Rented / Owned Media Experience • Trained Data and Computer Scientist with minor degree in psychology • Happy to have survived 10+ years in Advertising General Introduction What is Artificial Intelligence? A computer program which has the ability to think, solve problems, learn, set goals, and create. It is the science and engineering of making intelligent machines, especially intelligent computer programs. 4 - John McCarthy, 2002 Deep, Machine, Intelligence Learning? Deep Learning Machine Learning Artificial Intelligence Subarea of ML Subarea of AI Highly Simplified Where multi-layered Neural Algorithms that enable machines to A program that can “sense”, reason, Networks learn based on make decisions and improve by act, and adapt. high volumes of data learning over time. 5 Brief Timeline Deep Learning Machine Learning Artificial Intelligence 1965 1805-1957 Since 2000s - GAME CHANGER ”Birth”* of deep learning Algorithm groundwork and first self- Cost of computations and storage drop (Ukraine) learning algorithm * Principle of automatic models generation, inconclusive decisions, and selection by external criteria by Alexey Ivakhnenko 6 Keeping the Perspective That is what we mostly do Accomplishment of one or a very few goals NARROW AI 7 SUPERVISED Learning 8 UNSUPERVISED Learning 9 Some examples on how that works today What exactly is this about? What exactly is this about? It is about GO and MOVE 37 AI vs HI - Humans still are still able to win!!! A Major Challenge ExampleThe TRANSPARENCY on TRUST buildingProblem Even if the outcome is correct, the path to it might not necessarily be the one you expected.
    [Show full text]
  • A Depth of Deep Learning for Big Data and Its Applications
    Special Issue - 2020 International Journal of Engineering Research & Technology (IJERT) ISSN: 2278-0181 ENCADEMS - 2020 Conference Proceedings A Depth of Deep Learning for Big Data and its Applications Abhay Narayan Tripathi, Bharti Sharma Research Scholar, DIT Univeristy, Assistant Professor, DIT University Dehradun, Uttarakhand Dehradun,Uttrakhand ABSTRACT:- Although Machine Learning (ML) has become computational power and advances in data science. DL’s inherent synonymous for Artificial Intelligence (AI); recently, Deep ability to discover correlations from vast quantities of data in an Learning (DL) is being used in place of machine learning unsupervised fashion has been the main drive for its wide adoption. persistently. While machine learning is busy in supervised and Deep Learning also enables dynamic discovery of features from unsupervised methods, deep learning continues its motivation for data, unlike traditional machine learning approaches, where feature replicating the human nervous system by incorporating advanced selection remains a challenge. Deep Learning has been applied to types of Neural Networks (NN).. If we apply Deep Learning to various domains such as speech recognition and image Big Data, we can find unknown and useful patterns that classification, nature language processing, and computer vision. were impossible so far. Deep Learning is applied in self Typical deep neural networks (DNN) require large amounts of data driving cars, visual recognition, healthcare, transportation to learn parameters (often reaching to millions), which is a etc. Nowadays, companies have started to realize the importance computationally intensive process requiring significant time to train of data availability in large amounts in order to make the correct a model. As the data size increases exponentially and the deep decision and support their strategies.
    [Show full text]
  • Almond Variety Detection Using Deep Learning
    Almond Variety Detection using Deep Learning Benarous Ahmed Omar Farouq - a41369 Thesis presented to the School of Technology and Management in the scope of the Master in Information Systems. Supervisors: Prof. Maria João Tinoco Varanda Pereirar This document does not include the suggestions made by the board. Bragança 2019-2020 Dedication I dedicate my dissertation work to my family. A special feeling of gratitude to my loving parents, whose words of encouragement and push for tenacity ring in my ears. I also dedicate this dissertation to my many friends who have supported me throughout the process. I will always appreciate all they have done. v Acknowledgment In the name of Allah, the Most Gracious, the Most Merciful I would like to express my deep sense of gratitude, respect and heartfelt thanks to Prof. Maria João Tinoco Varanda Pereira, for her great contribu- tions, encouragement and support during the thesis writing, understanding and patience that have accompanied me during the research time. I would like to acknowledge and thank to ESTIG-IPB school division for providing me of the necessary materials needed to conduct my research and providing any assistance requested. I would like to acknowledge and thank to ESA-IPB school division for allow- ing me to conduct my research and providing any assistance requested. Special thanks goes to the members of the lab Professors Nuno Rodrigues and José Alberto Pereira for their continued support. Finally I would like to thank all professor of the hassiba benbouali university where i have started my journey of studying computer science. And special thank and acknowledgment to all community of computer science, this dynamic and healthy community is the reason why computer science, which is very new vi field compared to other sciences, has evolved so dramatically fast and now have huge impact on almost everything.
    [Show full text]
  • Deep Learning: State of the Art (2020) Deep Learning Lecture Series
    Deep Learning: State of the Art (2020) Deep Learning Lecture Series For the full list of references visit: http://bit.ly/deeplearn-sota-2020 https://deeplearning.mit.edu 2020 Outline • Deep Learning Growth, Celebrations, and Limitations • Deep Learning and Deep RL Frameworks • Natural Language Processing • Deep RL and Self-Play • Science of Deep Learning and Interesting Directions • Autonomous Vehicles and AI-Assisted Driving • Government, Politics, Policy • Courses, Tutorials, Books • General Hopes for 2020 For the full list of references visit: http://bit.ly/deeplearn-sota-2020 https://deeplearning.mit.edu 2020 “AI began with an ancient wish to forge the gods.” - Pamela McCorduck, Machines Who Think, 1979 Frankenstein (1818) Ex Machina (2015) Visualized here are 3% of the neurons and 0.0001% of the synapses in the brain. Thalamocortical system visualization via DigiCortex Engine. For the full list of references visit: http://bit.ly/deeplearn-sota-2020 https://deeplearning.mit.edu 2020 Deep Learning & AI in Context of Human History We are here Perspective: • Universe created 13.8 billion years ago • Earth created 4.54 billion years ago • Modern humans 300,000 years ago 1700s and beyond: Industrial revolution, steam • Civilization engine, mechanized factory systems, machine tools 12,000 years ago • Written record 5,000 years ago For the full list of references visit: http://bit.ly/deeplearn-sota-2020 https://deeplearning.mit.edu 2020 Artificial Intelligence in Context of Human History We are here Perspective: • Universe created 13.8 billion years ago • Earth created 4.54 billion years ago • Modern humans Dreams, mathematical foundations, and engineering in reality. 300,000 years ago Alan Turing, 1951: “It seems probable that once the machine • Civilization thinking method had started, it would not take long to outstrip 12,000 years ago our feeble powers.
    [Show full text]
  • Skoltech Annual Report 2019
    Annual Report Contents FOREWORD 002 YEAR IN NUMBERS 005 01 / FOCUS & GOVERNANCE 008 Human Capital 010 Institutional Governance & Development 030 Strategic Communications 038 02 / ACADEMIC & TECHNOLOGY EXCELLENCE 058 Teaching & Learning 062 Research 072 Integrating Innovation 088 03 / VALUE GENERATION 094 Industry Funded Research 096 New Enterprises & Technology Licensing 100 Professional Training & Advisory Services 102 04 / OPERATIONAL MANAGEMENT 108 GLOSSARY 112 1 It is my pleasure and a privilege to present international high-tech companies, research the annual report. Established in 2011 with and innovation agencies, in total of 1.6 bln the vision of being a world class university, rubles – 45% growth compared to 2018. Skoltech proudly marks 2019 as the year of Over 2 bln rubles of sponsored research success. funding have been confirmed for 2020 2019 was the first year in the new and beyond. Long-term investments have campus. Awarded at UNESCO headquarters been secured by Skoltech joint laboratories with the Prix Versailles as the world's best with Oerlikon, Gazprom Neft, Huawei, campus in 2019, it is without doubt an Sberbank, and Topcon. Also, Sberbank joined architectural masterpiece. At the same time, Skoltech as a founder. I am confident that the most important aspect of the campus this will bring new technological tasks and is its unique atmosphere and infrastructure opportunities for student internships and which inspire and enable our commitment for graduates’ employment. excellence. Staffed with top-notch professionals, the The high research performance of the Center of the Internet of Things contributes faculty, researchers, engineers, and students to the National Technology Initiative. The has put Skoltech into the top-100 Nature 5G Open Radio Access Network Center, Index 2019 Young Universities.
    [Show full text]
  • A Timeline of Artificial Intelligence
    A Timeline of Artificial Intelligence by piero scaruffi | www.scaruffi.com All of these events are explained in my book "Intelligence is not Artificial". TM, ®, Copyright © 1996-2019 Piero Scaruffi except pictures. All rights reserved. 1960: Henry Kelley and Arthur Bryson invent backpropagation 1960: Donald Michie's reinforcement-learning system MENACE 1960: Hilary Putnam's Computational Functionalism ("Minds and Machines") 1960: The backpropagation algorithm 1961: Melvin Maron's "Automatic Indexing" 1961: Karl Steinbuch's neural network Lernmatrix 1961: Leonard Scheer's and John Chubbuck's Mod I (1962) and Mod II (1964) 1961: Space General Corporation's lunar explorer 1962: IBM's "Shoebox" for speech recognition 1962: AMF's "VersaTran" robot 1963: John McCarthy moves to Stanford and founds the Stanford Artificial Intelligence Laboratory (SAIL) 1963: Lawrence Roberts' "Machine Perception of Three Dimensional Solids", the birth of computer vision 1963: Jim Slagle writes a program for symbolic integration (calculus) 1963: Edward Feigenbaum's and Julian Feldman's "Computers and Thought" 1963: Vladimir Vapnik's "support-vector networks" (SVN) 1964: Peter Toma demonstrates the machine-translation system Systran 1965: Irving John Good (Isidore Jacob Gudak) speculates about "ultraintelligent machines" (the "singularity") 1965: The Case Institute of Technology builds the first computer-controlled robotic arm 1965: Ed Feigenbaum's Dendral expert system 1965: Gordon Moore's Law of exponential progress in integrated circuits ("Cramming more components
    [Show full text]
  • Neurones, Les Intelligences Simulées
    Centre Pompidou Centre Dossier de presse Neurones, Direction de la communication et du numérique centrepompidou.fr les intelligences simulées 26 février – 20 avril 2020 Dans le cadre de Mutations / Créations #4 Neurones, les intelligences simulées 26 février – 20 avril 2020 Galerie 4, Niveau 1 Centre Pompidou Centre Dans le cadre de Mutations / Créations #4 Dossier Sommaire de presse p. 3 Direction de la communication Neurones, les intelligences simulées etDirection du numérique de la communication À propos de l’exposition et du numérique 75191 Paris cedex 04 centrepompidou.fr Démystifier l’intelligence artificielle p. 4 Directrice Agnès Benayer Entretien avec Frédéric Migayrou T. 00 33 (0)1 44 78 12 87 [email protected] Neurones, les intelligences simulées p. 5 Attaché de presse Le parcours de l’exposition Timothée Nicot Artistes exposés T. 00 33 (0)1 44 78 45 79 [email protected] Section 1 p. 6 - 7 centrepompidou.fr L’objet-cerveau #ExpoNeurones #MutationsCreations Section 2 p. 8 - 9 L’intelligence des jeux Section 3 p. 10 - 12 Cyberzoo Section 4 p. 13 - 16 Consciences augmentées Section 5 p. 17 - 19 Arbres et réseaux Publication p. 20 Informations pratiques p. 21 Dossier de presse Dossier de presse Neurones, les intelligences simulées les intelligences simulées Neurones Neurones 26 février – 20 avril 2020 Galerie 4, Niveau 1 À l’heure où l’intelligence artificielle s’étend à tous les domaines du monde contemporain, le Centre Pompidou propose pour la première Commissaires fois, avec « Neurones, les intelligences simulées », une mise en Frédéric Migayrou, Directeur adjoint du Musée national d’art relation de ce phénomène avec l’histoire des neurosciences et de la moderne-Centre de Création Industrielle, neuro-computation.
    [Show full text]
  • Deep Learning : ! Nouveaux Apports!
    Deep Learning : ! nouveaux apports! Christian Wolf! Université de Lyon, INSA-Lyon! LIRIS UMR CNRS 5205! 10 Mars, 2016! 1! 2! Data interpretation! ="?" 3! Friends and foes! 4! Where am I?! Visual landmarks! 5! « What is a cat? »! ! Children automatically learn to recognize objects from shapes and visual input given examples.! Human learning is mixed supervised / unsupervised! 6! Learning to predict! {dog, cat, pokemon, helicopter, …} {A. Turing, A. Einstein, S. Cooper, …} {0, 1, … 24, 25, 26, …, 98, 99, …} We would like to predict a value t from an observed input ! ! ! Parameters are learned from training data.! 7! A biological neuron ! Devin K. Phillips! 8! Recognition of complex activities through unsupervised learning of structured attribute models April 30, 2015 Abstract Complex human actionsArtificial are modelled as Neural graphs over basic networks attributes and! their spatial and temporal relationships. The attribute dictionary as well as the graphical structure are learned automatically from training« data. Perceptron »! D y(x, w)= wixi Xi=0 1 The activity model A set = c ,c ,...,c of C complex activities is considered, where each activity c is C { i 2 C } i modelled as a Graph i =(Vi,Ei), and both, nodes and edges of the graphs, are valued. For convenience, in the followingG we will drop the index i and consider a single graph for a given activity. The set of nodes V of the graph is indexed (here denoted by j)andcorrespondsto occurrences of basic attributes. EachG node j is assigned a vector of 4 values a ,x ,y ,t : { j j j j} the attribute type aj and a triplet of spatio-temporal coordinates xj,yj and tj.
    [Show full text]
  • Download Mobile App
    www.european-mrs.com CONFERENCE PROGRAMME 16 - 19 September 2019 Fall Meeting E-MRS CONFERENCE PROGRAMME - FALL MEETING 2019 Technical editor: Agnieszka Rytel Cover design and maps: Jarosław Pura General layout and DTP: Małgorzata Zielińska www.european-mrs.com E-MRS Fall Meeting Office: Faculty of Materials Science and Engineering Warsaw University of Technology ul. Wołoska 141, 02-507 Warsaw +48 22 234 87 94 [email protected] th th 16 – September 2019 Welcome We are pleased to welcome you to the E-MRS 2019 Fall Meeting. The European Materials Research Society (E-MRS) was established in 1983 through the initiative of individual european materials scientists. A number of european materials scientists who attended the MRS meetings in the U.S.A. realised that such a society could be of benefit to Europe to enhance the links between materials sci-ence and industry and to provide a voice for the materials community. Both the E-MRS Spring and Fall Meetings provide an international forum to discuss recent advances in the field of materials science. This year the conference is again being held in the Central Campus of the Warsaw University of Technology and is the 18th E-MRS Fall Meeting to be held in Warsaw since 2002. The conference will consist of: • 22 parallel symposia, • A plenary session which includes the presentation of the Jan Czochralski Award, • An Exhibition of products and services of interest to the conference participants. - 3 - Warsaw University of Technology Campus Overview - 4 - 16 th – 19 th September 2019 Main Building - 5
    [Show full text]
  • Deep Learning for Computer Graphics Introduction / Overview
    CS 6501: Deep Learning for Computer Graphics Introduction / Overview Connelly Barnes Outline • Course info: • Website • Goals / topics • Grading • Sign up sheet • Background: machine learning, neural networks, computer vision • Recent impact on computer graphics • Examples of papers we will cover Course Website http://www.connellybarnes.com/work/class/2016/deep_learning_graphics/ • Other ways to find the website: • Linked to from UVa Collab • Linked to from Lou’s list • Linked to from my website connellybarnes.com Course Goals • Learn math and computer science behind deep (artificial) neural networks • Understand in a general way how deep learning has been applied to areas such as natural language processing, robotics, games, finance • Understand in moderate depth how deep learning is used in computer vision • Understand in greater depth how deep learning is used in graphics • Gain experience building your own deep networks for visual computing problems (computer graphics, computer vision) Course Topics (More Detail) • Neural networks • Fully connected • Convolutional • Recurrent • Autoencoders • How to initialize • How to train • How to prevent problems: e.g. vanishing/exploding gradients Grading • Quizzes in class (15%) • Student paper presentations (20%) • Programming assignments (35%) • Final course project (30%) • No mid-term Sign Up Sheet • Everyone sign up for student presentation of a paper • Also enter your username if you do not have CS department account • Obtain sign up sheet URL from Collab. What is Machine Learning? • “The field of study that gives computers the ability to learn without being explicitly programmed.” – Arthur Samuel What is Machine Learning? • 3 broad categories: • Supervised learning: computer presented with example inputs and desired outputs by a “teacher”, goal is to learn general rule that maps inputs to outputs.
    [Show full text]
  • Multivariate Time Series Pattern Recognition Using Machine
    MULTIVARIATE TIME SERIES PATTERN RECOGNITION USING MACHINE LEARNING AND DEEP LEARNING METHODS Thesis Submitted to the Graduate Faculty of The University of Texas at Arlington in partial fulfillment of the requirements of the degree of Master of Science in Industrial Engineering By Sai Abhishek Devar B.Tech. Manipal Institute of Technology, 2016 December 2019 i Copyright @ by Sai Abhishek Devar 2019 All Rights Reserved ii Acknowledgments I would like to thank my advisor, Dr. Shouyi Wang, for his invaluable guidance and motivation throughout this work. I would also like to thank Dr. Kan Chen and Dr. Yuan Zhou for serving on my thesis defense committee. A special thanks to my family and friends for their unconditional love and support. iii Table of Content Acknowledgments .................................................................................................... iii Abstract .................................................................................................................... ix Chapter-1 .................................................................................................................... 1 Introduction ............................................................................................................. 1 1.1 Multivariate time series data ............................................................................. 2 1.2 Literature Review of Machine-learning & Deep-learning ............................... 2 Chapter-2 ...................................................................................................................
    [Show full text]