Learning Activation Functions in Deep Neural Networks

Total Page:16

File Type:pdf, Size:1020Kb

Learning Activation Functions in Deep Neural Networks UNIVERSITÉ DE MONTRÉAL LEARNING ACTIVATION FUNCTIONS IN DEEP NEURAL NETWORKS FARNOUSH FARHADI DÉPARTEMENT DE MATHÉMATIQUES ET DE GÉNIE INDUSTRIEL ÉCOLE POLYTECHNIQUE DE MONTRÉAL MÉMOIRE PRÉSENTÉ EN VUE DE L’OBTENTION DU DIPLÔME DE MAÎTRISE ÈS SCIENCES APPLIQUÉES (GÉNIE INDUSTRIEL) DÉCEMBRE 2017 c Farnoush Farhadi, 2017. UNIVERSITÉ DE MONTRÉAL ÉCOLE POLYTECHNIQUE DE MONTRÉAL Ce mémoire intitulé : LEARNING ACTIVATION FUNCTIONS IN DEEP NEURAL NETWORKS présenté par : FARHADI Farnoush en vue de l’obtention du diplôme de : Maîtrise ès sciences appliquées a été dûment accepté par le jury d’examen constitué de : M. ADJENGUE Luc-Désiré, Ph. D., président M. LODI Andrea, Ph. D., membre et directeur de recherche M. PARTOVI NIA Vahid, Doctorat, membre et codirecteur de recherche M. CHARLIN Laurent, Ph. D., membre iii DEDICATION This thesis is dedicated to my beloved parents, Ahmadreza and Sholeh, who are my first teachers and always love me unconditionally. This work is also dedicated to my love, Arash, who has been a great source of motivation and encouragement during the challenges of graduate studies and life. iv ACKNOWLEDGEMENTS I would like to express my gratitude to Prof. Andrea Lodi for his permission to be my supervisor at Ecole Polytechnique de Montreal and more importantly for his enthusiastic encouragements and invaluable continuous support during my research and education. I am very appreciated to him for introducing me to a MITACS internship in which I have developed myself both academically and professionally. I would express my deepest thanks to Dr. Vahid Partovi Nia, my co-supervisor at Ecole Poly- technique de Montreal, for his supportive and careful guidance and taking part in important decisions which were extremely precious for my research both theoretically and practically. I would also like to thank the examining committee for taking time out from their busy schedule to read and review my thesis and provide me with their insightful comments. Further acknowledgements go to the members of the Excellent Research Chair in Data Science for Real-time Decision-making who were eager to help me whenever needed. I would like to express my deepest gratitude and special thanks to Lane Cochrane, the Chief Analytics Officer at iPerceptions, for giving me the opportunity to do an internship within the company. For me it was a unique experience to be at iPerceptions and to do the research on machine learning. I would like to thank Sylvain Laporte, the Chief Technology Officer at iPerceptions, who in spite of being extraordinarily busy with his duties, took time out to give me necessary advice. I am especially thankful to all my collegues at R&D team at iPerceptions for their careful and valuable guidance to make life easier. And on a personal note, a special word of thanks goes to my parents who always stand behind me and support me spiritually throughout my life. My special thanks to Arshin, Mehrnoush and Siavash, my sisters and brother-in-law whose encourage and guides always help me in my personal and professional life. I am grateful for the love, encouragement, and tolerance of Arash, the man who has made all the difference in my life. Without his patience and love, I could not have completed this thesis. I am especially grateful to Atefeh, my best friend, who has been there whenever I needed her. She always has the time to hear about my successes and failures and has always been a faithful listener. I would like to thank my dear friend David Berger for helping me in translating the abstract into French and Chris Nael for his assistance in general editing. v RÉSUMÉ Récemment, la prédiction de l’intention des utilisateurs basée sur les données comportemen- tales est devenue plus attrayante dans les domaines de marketing numérique et publicité en ligne. La prédiction est souvent effectuée par analyse de l’information du comportement des visiteurs qui comprend des détails sur la façon dont chaque utilisateur a visité le site Web commercial. Dans ce travail, nous explorons les méthodes statistiques d’apprentissage auto- matique pour classifier les visiteurs d’un site Web de commerce électronique d’après des URL qu’ils ont parcourues. Plus précisément, on vise à étudier comment les réseaux de neurones profonds de pointe pourraient efficacement prédire le but des visiteurs dans une session de visite, en utilisant uniquement des séquences d’URL. Dans un réseau contenant des couches limitées, le choix de la fonction d’activation a un effet sur l’apprentissage de la représentation et les performances du réseau. Typiquement, les modèles de réseaux de neurones sont appris en utilisant des fonctions d’ac- tivation spécifiques comme la fonction sigmoïde symétrique standard ou relu linéaire. Dans cette mémoire, on vise à proposer une fonction sigmoïde asymetrique paramétrée dont la paramètre pourrait être controlé et ajusté dans chacun des neurons cachées avec d’autres paramètres de réseau profond en formation. De plus, nous visons à proposer une variante non-linéaire de la fonction relu pour les réseaux profonds. Comme le sigmoïde adaptatif, notre objectif est de régler les paramètres du relu adaptatif pour chaque unité individuelle- ment, dans l’étape de formation. Des méthodes et des algorithmes pour développer ces fonctions d’activation adaptatives sont discutés. En outre, une petite variante de MLP (Multi Layer Perceptron) et un modèle CNN (Convolutional Neural Network) applicant nos fonctions d’activation proposées sont utilisés pour prédire l’intention des utilisateurs selon les données d’URL. Quatre jeux de données différents ont été choisis, appelé les données simulées, les données MNIST, les données de revue de film, et les données d’URL pour démontrer l’effet de sélectionner différentes fonctions d’activation sur les modèles MLP et CNN proposés. vi ABSTRACT Recently, user intention prediction based on behavioral data has become more attractive in digital marketing and online advertisement. The prediction is often performed by analyzing the logged information which includes details on how each user visited the commercial web- site. In this work, we explore the machine learning methods for classification of visitors to an e-commerce website based on URLs they visited. We aim at studying on how effectively the state-of-the-art deep neural networks could predict the purpose of visitors in a session using only URL sequences. Typical deep neural networks employ a fixed nonlinear activation function for each hidden neuron. In this thesis, two adaptive activation functions for individ- ual hidden units are proposed such that the deep network learns these activation functions in training. Methods and algorithms for developing these adaptive activation functions are discussed. Furthermore, performance of the proposed activation functions in deep networks compared to state-of-the-art models are also evaluated on a real-world URL dataset and two well-known benchmarks, MNIST and Movie Review benchmarks as well. vii TABLE OF CONTENTS DEDICATION . iii ACKNOWLEDGEMENTS . iv RÉSUMÉ . v ABSTRACT . vi TABLE OF CONTENTS . vii LIST OF TABLES . x LIST OF FIGURES . xii LIST OF ABBREVIATIONS AND NOTATIONS . xvii LIST OF APPENDICES . xx CHAPTER 1 INTRODUCTION . 1 1.1 Deep Neural Networks . .3 1.2 Adaptive Activation Functions . .5 1.3 Research Objectives . .6 1.4 Our Contributions . .7 1.5 Thesis Structure . .7 CHAPTER 2 OVERVIEW . 8 2.1 Linear Regression . .8 2.1.1 Maximum Likelihood Estimation and Least Squares . 10 2.1.2 Multiple Outputs . 12 2.2 Generalized Linear Models (GLMs) . 13 2.2.1 Exponential Family . 14 2.2.2 Exponential Dispersion Family . 15 2.2.3 Mean and Variance . 16 2.2.4 Linear Structure . 18 2.2.5 GLMs Components . 19 2.2.6 Maximum Likelihood Estimates . 21 viii 2.2.7 Logistic Regression . 25 2.2.8 Maximum Likelihood Estimates in Logistic Regression . 27 CHAPTER 3 DEEP NEURAL NETWORKS . 33 3.1 Perceptron . 33 3.1.1 Perceptron Algorithm . 33 3.1.2 Model . 33 3.1.3 Basic Perceptron Learning Rule and Convergence . 36 3.1.4 Perceptron Learning Rule on Loss Function . 42 3.1.5 Perceptron with Sigmoid Function . 45 3.1.6 Perceptron Issues . 51 3.2 Multi Layer Perceptrons (MLPs) . 52 3.2.1 Activation Functions . 56 3.2.2 Loss Function . 59 3.2.3 Backpropagation Learning Algorithm . 60 3.2.4 Mini-batch Stochastic Gradient Descent (Mini-batch SGD) . 64 3.2.5 Regularization . 66 3.2.6 Dropout . 67 3.2.7 Parameters Initialization . 68 3.3 Convolutional Neural Networks (CNNs) . 68 3.3.1 Convolutional Layer . 69 3.3.2 Max Pooling Layer . 71 3.3.3 Loss Function and Gradients . 72 CHAPTER 4 ADAPTIVE ACTIVATION FUNCTIONS . 74 4.1 Latent Variable Formulation . 74 4.2 The complementary log-log transformation . 77 4.2.1 Maximum Likelihood Estimates with Complementary log-log . 79 4.3 Adaptive Sigmoid Link Function . 81 4.3.1 Identifiability . 83 4.3.2 Maximum Likelihood Estimates for Adaptive Sigmoid Link Function 86 4.3.3 ANNs with Adaptive-sigmoid : Loss Function and Optimization . 87 4.4 Adaptive ReLu Function . 90 4.4.1 ANNs with Adaptive-relu : Loss Function and Optimization . 92 CHAPTER 5 EXPERIMENTS AND RESULTS . 95 5.1 Performance Measures . 95 ix 5.2 Experiment 1 : Simulated Data . 97 5.2.1 Data . 98 5.2.2 Tools and Technologies . 98 5.2.3 Results and Evaluations . 98 5.2.4 Conclusions . 108 5.3 Experiment 2 : MNIST Benchmark . 110 5.3.1 data . 110 5.3.2 Tools and Technologies . 110 5.3.3 Models . 110 5.3.4 Results and Evaluations . 111 5.3.5 Conclusions . 114 5.4 Experiment 3 : Sentiment Analysis . 117 5.4.1 Data . 117 5.4.2 Data preparation . 117 5.4.3 Models . 118 5.4.4 Tools and Technologies .
Recommended publications
  • 6.5 Applications of Exponential and Logarithmic Functions 469
    6.5 Applications of Exponential and Logarithmic Functions 469 6.5 Applications of Exponential and Logarithmic Functions As we mentioned in Section 6.1, exponential and logarithmic functions are used to model a wide variety of behaviors in the real world. In the examples that follow, note that while the applications are drawn from many different disciplines, the mathematics remains essentially the same. Due to the applied nature of the problems we will examine in this section, the calculator is often used to express our answers as decimal approximations. 6.5.1 Applications of Exponential Functions Perhaps the most well-known application of exponential functions comes from the financial world. Suppose you have $100 to invest at your local bank and they are offering a whopping 5 % annual percentage interest rate. This means that after one year, the bank will pay you 5% of that $100, or $100(0:05) = $5 in interest, so you now have $105.1 This is in accordance with the formula for simple interest which you have undoubtedly run across at some point before. Equation 6.1. Simple Interest The amount of interest I accrued at an annual rate r on an investmenta P after t years is I = P rt The amount A in the account after t years is given by A = P + I = P + P rt = P (1 + rt) aCalled the principal Suppose, however, that six months into the year, you hear of a better deal at a rival bank.2 Naturally, you withdraw your money and try to invest it at the higher rate there.
    [Show full text]
  • Chapter 10. Logistic Models
    www.ck12.org Chapter 10. Logistic Models CHAPTER 10 Logistic Models Chapter Outline 10.1 LOGISTIC GROWTH MODEL 10.2 CONDITIONS FAVORING LOGISTIC GROWTH 10.3 REFERENCES 207 10.1. Logistic Growth Model www.ck12.org 10.1 Logistic Growth Model Here you will explore the graph and equation of the logistic function. Learning Objectives • Recognize logistic functions. • Identify the carrying capacity and inflection point of a logistic function. Logistic Models Exponential growth increases without bound. This is reasonable for some situations; however, for populations there is usually some type of upper bound. This can be caused by limitations on food, space or other scarce resources. The effect of this limiting upper bound is a curve that grows exponentially at first and then slows down and hardly grows at all. This is characteristic of a logistic growth model. The logistic equation is of the form: C C f (t) = 1+ab−t = 1+ae−kt The above equations represent the logistic function, and it contains three important pieces: C, a, and b. C determines that maximum value of the function, also known as the carrying capacity. C is represented by the dashed line in the graph below. 208 www.ck12.org Chapter 10. Logistic Models The constant of a in the logistic function is used much like a in the exponential function: it helps determine the value of the function at t=0. Specifically: C f (0) = 1+a The constant of b also follows a similar concept to the exponential function: it helps dictate the rate of change at the beginning and the end of the function.
    [Show full text]
  • Neural Network Architectures and Activation Functions: a Gaussian Process Approach
    Fakultät für Informatik Neural Network Architectures and Activation Functions: A Gaussian Process Approach Sebastian Urban Vollständiger Abdruck der von der Fakultät für Informatik der Technischen Universität München zur Erlangung des akademischen Grades eines Doktors der Naturwissenschaften (Dr. rer. nat.) genehmigten Dissertation. Vorsitzender: Prof. Dr. rer. nat. Stephan Günnemann Prüfende der Dissertation: 1. Prof. Dr. Patrick van der Smagt 2. Prof. Dr. rer. nat. Daniel Cremers 3. Prof. Dr. Bernd Bischl, Ludwig-Maximilians-Universität München Die Dissertation wurde am 22.11.2017 bei der Technischen Universtät München eingereicht und durch die Fakultät für Informatik am 14.05.2018 angenommen. Neural Network Architectures and Activation Functions: A Gaussian Process Approach Sebastian Urban Technical University Munich 2017 ii Abstract The success of applying neural networks crucially depends on the network architecture being appropriate for the task. Determining the right architecture is a computationally intensive process, requiring many trials with different candidate architectures. We show that the neural activation function, if allowed to individually change for each neuron, can implicitly control many aspects of the network architecture, such as effective number of layers, effective number of neurons in a layer, skip connections and whether a neuron is additive or multiplicative. Motivated by this observation we propose stochastic, non-parametric activation functions that are fully learnable and individual to each neuron. Complexity and the risk of overfitting are controlled by placing a Gaussian process prior over these functions. The result is the Gaussian process neuron, a probabilistic unit that can be used as the basic building block for probabilistic graphical models that resemble the structure of neural networks.
    [Show full text]
  • Neural Networks
    School of Computer Science 10-601B Introduction to Machine Learning Neural Networks Readings: Matt Gormley Bishop Ch. 5 Lecture 15 Murphy Ch. 16.5, Ch. 28 October 19, 2016 Mitchell Ch. 4 1 Reminders 2 Outline • Logistic Regression (Recap) • Neural Networks • Backpropagation 3 RECALL: LOGISTIC REGRESSION 4 Using gradient ascent for linear Recall… classifiers Key idea behind today’s lecture: 1. Define a linear classifier (logistic regression) 2. Define an objective function (likelihood) 3. Optimize it with gradient descent to learn parameters 4. Predict the class with highest probability under the model 5 Using gradient ascent for linear Recall… classifiers This decision function isn’t Use a differentiable function differentiable: instead: 1 T pθ(y =1t)= h(t)=sign(θ t) | 1+2tT( θT t) − sign(x) 1 logistic(u) ≡ −u 1+ e 6 Using gradient ascent for linear Recall… classifiers This decision function isn’t Use a differentiable function differentiable: instead: 1 T pθ(y =1t)= h(t)=sign(θ t) | 1+2tT( θT t) − sign(x) 1 logistic(u) ≡ −u 1+ e 7 Recall… Logistic Regression Data: Inputs are continuous vectors of length K. Outputs are discrete. (i) (i) N K = t ,y where t R and y 0, 1 D { }i=1 ∈ ∈ { } Model: Logistic function applied to dot product of parameters with input vector. 1 pθ(y =1t)= | 1+2tT( θT t) − Learning: finds the parameters that minimize some objective function. θ∗ = argmin J(θ) θ Prediction: Output is the most probable class. yˆ = `;Kt pθ(y t) y 0,1 | ∈{ } 8 NEURAL NETWORKS 9 Learning highly non-linear functions f: X à Y l f might be non-linear
    [Show full text]
  • Lecture 5: Logistic Regression 1 MLE Derivation
    CSCI 5525 Machine Learning Fall 2019 Lecture 5: Logistic Regression Feb 10 2020 Lecturer: Steven Wu Scribe: Steven Wu Last lecture, we give several convex surrogate loss functions to replace the zero-one loss func- tion, which is NP-hard to optimize. Now let us look into one of the examples, logistic loss: given d parameter w and example (xi; yi) 2 R × {±1g, the logistic loss of w on example (xi; yi) is defined as | ln (1 + exp(−yiw xi)) This loss function is used in logistic regression. We will introduce the statistical model behind logistic regression, and show that the ERM problem for logistic regression is the same as the relevant maximum likelihood estimation (MLE) problem. 1 MLE Derivation For this derivation it is more convenient to have Y = f0; 1g. Note that for any label yi 2 f0; 1g, we also have the “signed” version of the label 2yi − 1 2 {−1; 1g. Recall that in general su- pervised learning setting, the learner receive examples (x1; y1);:::; (xn; yn) drawn iid from some distribution P over labeled examples. We will make the following parametric assumption on P : | yi j xi ∼ Bern(σ(w xi)) where Bern denotes the Bernoulli distribution, and σ is the logistic function defined as follows 1 exp(z) σ(z) = = 1 + exp(−z) 1 + exp(z) See Figure 1 for a visualization of the logistic function. In general, the logistic function is a useful function to convert real values into probabilities (in the range of (0; 1)). If w|x increases, then σ(w|x) also increases, and so does the probability of Y = 1.
    [Show full text]
  • An Introduction to Logistic Regression: from Basic Concepts to Interpretation with Particular Attention to Nursing Domain
    J Korean Acad Nurs Vol.43 No.2, 154 -164 J Korean Acad Nurs Vol.43 No.2 April 2013 http://dx.doi.org/10.4040/jkan.2013.43.2.154 An Introduction to Logistic Regression: From Basic Concepts to Interpretation with Particular Attention to Nursing Domain Park, Hyeoun-Ae College of Nursing and System Biomedical Informatics National Core Research Center, Seoul National University, Seoul, Korea Purpose: The purpose of this article is twofold: 1) introducing logistic regression (LR), a multivariable method for modeling the relationship between multiple independent variables and a categorical dependent variable, and 2) examining use and reporting of LR in the nursing literature. Methods: Text books on LR and research articles employing LR as main statistical analysis were reviewed. Twenty-three articles published between 2010 and 2011 in the Journal of Korean Academy of Nursing were analyzed for proper use and reporting of LR models. Results: Logistic regression from basic concepts such as odds, odds ratio, logit transformation and logistic curve, assumption, fitting, reporting and interpreting to cautions were presented. Substantial short- comings were found in both use of LR and reporting of results. For many studies, sample size was not sufficiently large to call into question the accuracy of the regression model. Additionally, only one study reported validation analysis. Conclusion: Nurs- ing researchers need to pay greater attention to guidelines concerning the use and reporting of LR models. Key words: Logit function, Maximum likelihood estimation, Odds, Odds ratio, Wald test INTRODUCTION The model serves two purposes: (1) it can predict the value of the depen- dent variable for new values of the independent variables, and (2) it can Multivariable methods of statistical analysis commonly appear in help describe the relative contribution of each independent variable to general health science literature (Bagley, White, & Golomb, 2001).
    [Show full text]
  • Fm 3-34 Engineer Operations Headquarters, Department
    FM 3-34 April 2009 ENGINEER OPERATIONS DISTRIBUTION RESTRICTION: Approved for public release; distribution is unlimited. HEADQUARTERS, DEPARTMENT OF THE ARMY This publication is available at Army Knowledge Online (AKO) <www.us.army.mil> and the General Dennis J. Reimer Training and Doctrine Digital Library at <www.train.army.mil>. *FM 3-34 Field Manual No. Headquarters 3-34 Department of the Army Washington, DC, 2 April 2009 Engineer Operations Contents Page PREFACE .............................................................................................................. v INTRODUCTION .................................................................................................. vii Chapter 1 THE OPERATIONAL ENVIRONMENT ............................................................. 1-1 Understanding the Operational Environment ..................................................... 1-1 The Military Variable ........................................................................................... 1-4 Spectrum of Requirements ................................................................................. 1-5 Support Spanning the Levels of War.................................................................. 1-7 Engineer Soldiers ............................................................................................... 1-9 Chapter 2 ENGINEERING IN UNIFIED ACTION ............................................................... 2-1 SECTION I–THE ENGINEER REGIMENT ........................................................ 2-1 The
    [Show full text]
  • Notes: Logistic Functions I. Use, Function, and General Graph If
    Notes: Logistic Functions I. Use, Function, and General Graph If growth begins slowly, then increases rapidly and eventually levels off, the data often can be model by an “S-curve”, or a logistic function. The logistic function describes certain kinds of growth. These functions, like exponential functions, grow quickly at first, but because of restrictions that place limits on the size of the underlying population, eventually grow more slowly and then level off. For real numbers a, b, and c, the function is a logistic function. If a >0, a logistic function increases when b >0 and decreases when b < 0. Logistic Growth Function Equation: x-Intercept: none y-intercept: Horizontal Asymptote: The number, , is called the limiting value or the upper limit of the function because the graph of a logistic growth function will have a horizontal asymptote at y = c. As is clear from the graph above, the characteristic S-shape in the graph of a logistic function shows that initial exponential growth is followed by a period in which growth slows and then levels off, approaching (but never attaining) a maximum upper limit. Logistic functions are good models of biological population growth in species which have grown so large that they are near to saturating their ecosystems, or of the spread of information within societies. They are also common in marketing, where they chart the sales of new products over time; in a different context, they can also describe demand curves: the decline of demand for a product as a function of increasing price can be modeled by a logistic function, as in the figure below.
    [Show full text]
  • Sigmoid Functions in Reliability Based Management 2007 15 2 67 2.1 the Nature of Performance Growth
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Periodica Polytechnica (Budapest University of Technology and Economics) Ŕ periodica polytechnica Sigmoid functions in reliability based Social and Management Sciences management 15/2 (2007) 67–72 doi: 10.3311/pp.so.2007-2.04 Tamás Jónás web: http://www.pp.bme.hu/so c Periodica Polytechnica 2007 RESEARCH ARTICLE Received 2008-09-14 Abstract 1 Introduction This paper introduces the theoretical background of a few A sigmoid function is a mathematical function that produces possible applications of the so-called sigmoid-type functions in a curve having an "S" shape, and is defined by the manufacturing and service management. 1 σλ,x (x) (1) An extended concept of reliability, derived from fuzzy theory, 0 λ(x x0) = 1 e− − is discussed here to illustrate how reliability based management + formula. The sigmoid function is also called sigmoidal curve decisions can be made consistent, when handling of weakly de- [1] or logistic function. The interpretation of sigmoid-type func- fined concepts is needed. tions – I use here – is that any function that can be transformed I demonstrate how performance growth can be modelled us- into σλ,x (x) through substitutions and linear transformations ing an aggregate approach with support of sigmoid-type func- 0 is considered as a sigmoid-type one. There are several well- tions. known applications of sigmoid-type functions. A few examples Suitably chosen parameters of sigmoid-type functions allow are: threshold function in neural networks [2], approximation these to be used as failure probability distribution and survival of Gaussian probability distribution, logistic regression [3], or functions.
    [Show full text]
  • 2.161 Signal Processing: Continuous and Discrete Fall 2008
    MIT OpenCourseWare http://ocw.mit.edu 2.161 Signal Processing: Continuous and Discrete Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. MASSACHUSETTS INSTITUTE OF TECHNOLOGY DEPARTMENT OF MECHANICAL ENGINEERING 2.161 Signal Processing – Continuous and Discrete 1 The Laplace Transform 1 Introduction In the class handout Introduction to Frequency Domain Processing we introduced the Fourier trans­ form as an important theoretical and practical tool for the analysis of waveforms, and the design of linear filters. We noted that there are classes of waveforms for which the classical Fourier integral does not converge. An important function that does not have a classical Fourier transform are is the unit step (Heaviside) function ( 0 t · 0; us(t) = 1 t > 0; Clearly Z 1 jus(t)j dt = 1; ¡1 and the forward Fourier integral Z 1 Z 1 ¡j­t ¡j­t Us(jΩ) = us(t)e dt = e dt (1) ¡1 0 does not converge. Aside: We also saw in the handout that many such functions whose Fourier integrals do not converge do in fact have Fourier transforms that can be defined using distributions (with the use of the Dirac delta function ±(t)), for example we found that 1 F fu (t)g = ¼±(Ω) + : s jΩ Clearly any function f(t) in which limj tj!1 f(t) 6= 0 will have a Fourier integral that does not converge. Similarly, the ramp function ( 0 t · 0; r(t) = t t > 0: is not integrable in the absolute sense, and does not allow direct computation of the Fourier trans­ form.
    [Show full text]
  • Applied Machine Learning
    Applied Machine Learning Annalisa Marsico OWL RNA Bionformatics group Max Planck Institute for Molecular Genetics Free University of Berlin 13 Mai, SoSe 2015 Neural Networks Motivation: the non-linear hypothesis x x x Goal: learn the non-linear problem x x x x x One solution: add a lot of non-linear features x as input to the logistic function x x x x x x x x x = ( + + + + x + + + ⋯ . ) x x g = sigmoid function If more than two dimensions, including polynomial feature not a good idea It might lead to over fitting. Neural Networks • Solution : SVM with projection in a high-dimensional space. Kernel ‘trick’. • New solution: Neural Networks to learn non-linear complex hypothesis Neural Networks (NNs) – Biological motivation • Originally motivated by the desire of having machines which can mimic the brain • Widely used in the 80s and early 90s, popularity diminished in the late 90s • Recent resurgence: more computational power to run large-scale NNs The ‘brain’ algorithm • Fascinating hypothesis (Roe et al. 1992): the brain functions through a single learning algorithm Auditory cortex X Neurons Schematic representation of a single neuron in the brain - Neuron is a computational unit ‘Inputs’ (dendrites) Synapse Node of Cell body Ranvier Schwann cell Myelin sheath Cell nucleus Communication between neurons Modeling neurons Logistic unit Activation function is often a sigmoid 1 ≡ ℎ = ≡ () 1 + There exist other activation function: step function, linear function Logistic unit representation Bias unit, =1 () Sigmoid
    [Show full text]
  • Multilayer Perceptron
    Advanced Introduction to Machine Learning, CMU-10715 Perceptron, Multilayer Perceptron Barnabás Póczos, Sept 17 Contents History of Artificial Neural Networks Definitions: Perceptron, MLP Representation questions Perceptron algorithm Backpropagation algorithm 2 Short History Progression (1943-1960) • First mathematical model of neurons . Pitts & McCulloch (1943) • Beginning of artificial neural networks • Perceptron, Rosenblatt (1958) . A single layer neuron for classification . Perceptron learning rule . Perceptron convergence theorem Degression (1960-1980) • Perceptron can’t even learn the XOR function • We don’t know how to train MLP • 1969 Backpropagation… but not much attention… 3 Short History Progression (1980-) • 1986 Backpropagation reinvented: . Rumelhart, Hinton, Williams: Learning representations by back-propagating errors. Nature, 323, 533—536, 1986 • Successful applications: . Character recognition, autonomous cars,… • Open questions: Overfitting? Network structure? Neuron number? Layer number? Bad local minimum points? When to stop training? • Hopfield nets (1982), Boltzmann machines,… 4 Short History Degression (1993-) • SVM: Vapnik and his co-workers developed the Support Vector Machine (1993). It is a shallow architecture. • SVM almost kills the ANN research. • Training deeper networks consistently yields poor results. • Exception: deep convolutional neural networks, Yann LeCun 1998. (discriminative model) 5 Short History Progression (2006-) Deep Belief Networks (DBN) • Hinton, G. E, Osindero, S., and Teh, Y.
    [Show full text]