Natural-Logarithm-Rectified Activation Function in Convolutional Neural Networks
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Training Autoencoders by Alternating Minimization
Under review as a conference paper at ICLR 2018 TRAINING AUTOENCODERS BY ALTERNATING MINI- MIZATION Anonymous authors Paper under double-blind review ABSTRACT We present DANTE, a novel method for training neural networks, in particular autoencoders, using the alternating minimization principle. DANTE provides a distinct perspective in lieu of traditional gradient-based backpropagation techniques commonly used to train deep networks. It utilizes an adaptation of quasi-convex optimization techniques to cast autoencoder training as a bi-quasi-convex optimiza- tion problem. We show that for autoencoder configurations with both differentiable (e.g. sigmoid) and non-differentiable (e.g. ReLU) activation functions, we can perform the alternations very effectively. DANTE effortlessly extends to networks with multiple hidden layers and varying network configurations. In experiments on standard datasets, autoencoders trained using the proposed method were found to be very promising and competitive to traditional backpropagation techniques, both in terms of quality of solution, as well as training speed. 1 INTRODUCTION For much of the recent march of deep learning, gradient-based backpropagation methods, e.g. Stochastic Gradient Descent (SGD) and its variants, have been the mainstay of practitioners. The use of these methods, especially on vast amounts of data, has led to unprecedented progress in several areas of artificial intelligence. On one hand, the intense focus on these techniques has led to an intimate understanding of hardware requirements and code optimizations needed to execute these routines on large datasets in a scalable manner. Today, myriad off-the-shelf and highly optimized packages exist that can churn reasonably large datasets on GPU architectures with relatively mild human involvement and little bootstrap effort. -
Improvements on Activation Functions in ANN: an Overview
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by CSCanada.net: E-Journals (Canadian Academy of Oriental and Occidental Culture,... ISSN 1913-0341 [Print] Management Science and Engineering ISSN 1913-035X [Online] Vol. 14, No. 1, 2020, pp. 53-58 www.cscanada.net DOI:10.3968/11667 www.cscanada.org Improvements on Activation Functions in ANN: An Overview YANG Lexuan[a],* [a]Hangzhou No.2 High School of Zhejiang Province; Hangzhou, China. An ANN is comprised of computable units called * Corresponding author. neurons (or nodes). The activation function in a neuron processes the inputted signals and determines the output. Received 9 April 2020; accepted 21 May 2020 Published online 26 June 2020 Activation functions are particularly important in the ANN algorithm; change in activation functions can have a significant impact on network performance. Therefore, Abstract in recent years, researches have been done in depth on Activation functions are an essential part of artificial the improvement of activation functions to solve or neural networks. Over the years, researches have been alleviate some problems encountered in practices with done to seek for new functions that perform better. “classic” activation functions. This paper will give some There are several mainstream activation functions, such brief examples and discussions on the modifications and as sigmoid and ReLU, which are widely used across improvements of mainstream activation functions. the decades. At the meantime, many modified versions of these functions are also proposed by researchers in order to further improve the performance. In this paper, 1. INTRODUCTION OF ACTIVATION limitations of the mainstream activation functions, as well as the main characteristics and relative performances of FUNCTIONS their modifications, are discussed. -
Neural Network in Hardware
UNLV Theses, Dissertations, Professional Papers, and Capstones 12-15-2019 Neural Network in Hardware Jiong Si Follow this and additional works at: https://digitalscholarship.unlv.edu/thesesdissertations Part of the Electrical and Computer Engineering Commons Repository Citation Si, Jiong, "Neural Network in Hardware" (2019). UNLV Theses, Dissertations, Professional Papers, and Capstones. 3845. http://dx.doi.org/10.34917/18608784 This Dissertation is protected by copyright and/or related rights. It has been brought to you by Digital Scholarship@UNLV with permission from the rights-holder(s). You are free to use this Dissertation in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s) directly, unless additional rights are indicated by a Creative Commons license in the record and/or on the work itself. This Dissertation has been accepted for inclusion in UNLV Theses, Dissertations, Professional Papers, and Capstones by an authorized administrator of Digital Scholarship@UNLV. For more information, please contact [email protected]. NEURAL NETWORKS IN HARDWARE By Jiong Si Bachelor of Engineering – Automation Chongqing University of Science and Technology 2008 Master of Engineering – Precision Instrument and Machinery Hefei University of Technology 2011 A dissertation submitted in partial fulfillment of the requirements for the Doctor of Philosophy – Electrical Engineering Department of Electrical and Computer Engineering Howard R. Hughes College of Engineering The Graduate College University of Nevada, Las Vegas December 2019 Copyright 2019 by Jiong Si All Rights Reserved Dissertation Approval The Graduate College The University of Nevada, Las Vegas November 6, 2019 This dissertation prepared by Jiong Si entitled Neural Networks in Hardware is approved in partial fulfillment of the requirements for the degree of Doctor of Philosophy – Electrical Engineering Department of Electrical and Computer Engineering Sarah Harris, Ph.D. -
Population Dynamics: Variance and the Sigmoid Activation Function ⁎ André C
www.elsevier.com/locate/ynimg NeuroImage 42 (2008) 147–157 Technical Note Population dynamics: Variance and the sigmoid activation function ⁎ André C. Marreiros, Jean Daunizeau, Stefan J. Kiebel, and Karl J. Friston Wellcome Trust Centre for Neuroimaging, University College London, UK Received 24 January 2008; revised 8 April 2008; accepted 16 April 2008 Available online 29 April 2008 This paper demonstrates how the sigmoid activation function of (David et al., 2006a,b; Kiebel et al., 2006; Garrido et al., 2007; neural-mass models can be understood in terms of the variance or Moran et al., 2007). All these models embed key nonlinearities dispersion of neuronal states. We use this relationship to estimate the that characterise real neuronal interactions. The most prevalent probability density on hidden neuronal states, using non-invasive models are called neural-mass models and are generally for- electrophysiological (EEG) measures and dynamic casual modelling. mulated as a convolution of inputs to a neuronal ensemble or The importance of implicit variance in neuronal states for neural-mass population to produce an output. Critically, the outputs of one models of cortical dynamics is illustrated using both synthetic data and ensemble serve as input to another, after some static transforma- real EEG measurements of sensory evoked responses. © 2008 Elsevier Inc. All rights reserved. tion. Usually, the convolution operator is linear, whereas the transformation of outputs (e.g., mean depolarisation of pyramidal Keywords: Neural-mass models; Nonlinear; Modelling cells) to inputs (firing rates in presynaptic inputs) is a nonlinear sigmoidal function. This function generates the nonlinear beha- viours that are critical for modelling and understanding neuronal activity. -
Persian Handwritten Digit Recognition Using Combination of Convolutional Neural Network and Support Vector Machine Methods
572 The International Arab Journal of Information Technology, Vol. 17, No. 4, July 2020 Persian Handwritten Digit Recognition Using Combination of Convolutional Neural Network and Support Vector Machine Methods Mohammad Parseh, Mohammad Rahmanimanesh, and Parviz Keshavarzi Faculty of Electrical and Computer Engineering, Semnan University, Iran Abstract: Persian handwritten digit recognition is one of the important topics of image processing which significantly considered by researchers due to its many applications. The most important challenges in Persian handwritten digit recognition is the existence of various patterns in Persian digit writing that makes the feature extraction step to be more complicated.Since the handcraft feature extraction methods are complicated processes and their performance level are not stable, most of the recent studies have concentrated on proposing a suitable method for automatic feature extraction. In this paper, an automatic method based on machine learning is proposed for high-level feature extraction from Persian digit images by using Convolutional Neural Network (CNN). After that, a non-linear multi-class Support Vector Machine (SVM) classifier is used for data classification instead of fully connected layer in final layer of CNN. The proposed method has been applied to HODA dataset and obtained 99.56% of recognition rate. Experimental results are comparable with previous state-of-the-art methods. Keywords: Handwritten Digit Recognition, Convolutional Neural Network, Support Vector Machine. Received January 1, 2019; accepted November 11, 2019 https://doi.org/10.34028/iajit/17/4/16 1. Introduction years, they are good alternatives to handcraft feature extraction method. Optical Character Recognition (OCR) is one of the attractive topics of Artificial Intelligence [3, 6, 15, 23, 24]. -
Dynamic Modification of Activation Function Using the Backpropagation Algorithm in the Artificial Neural Networks
(IJACSA) International Journal of Advanced Computer Science and Applications, Vol. 10, No. 4, 2019 Dynamic Modification of Activation Function using the Backpropagation Algorithm in the Artificial Neural Networks Marina Adriana Mercioni1, Alexandru Tiron2, Stefan Holban3 Department of Computer Science Politehnica University Timisoara, Timisoara, Romania Abstract—The paper proposes the dynamic modification of These functions represent the main activation functions, the activation function in a learning technique, more exactly being currently the most used by neural networks, representing backpropagation algorithm. The modification consists in from this reason a standard in building of an architecture of changing slope of sigmoid function for activation function Neural Network type [6,7]. according to increase or decrease the error in an epoch of learning. The study was done using the Waikato Environment for The activation functions that have been proposed along the Knowledge Analysis (WEKA) platform to complete adding this time were analyzed in the light of the applicability of the feature in Multilayer Perceptron class. This study aims the backpropagation algorithm. It has noted that a function that dynamic modification of activation function has changed to enables differentiated rate calculated by the neural network to relative gradient error, also neural networks with hidden layers be differentiated so and the error of function becomes have not used for it. differentiated [8]. Keywords—Artificial neural networks; activation function; The main purpose of activation function consists in the sigmoid function; WEKA; multilayer perceptron; instance; scalation of outputs of the neurons in neural networks and the classifier; gradient; rate metric; performance; dynamic introduction a non-linear relationship between the input and modification output of the neuron [9]. -
Two-Steps Implementation of Sigmoid Function for Artificial Neural Network in Field Programmable Gate Array
VOL. 11, NO. 7, APRIL 2016 ISSN 1819-6608 ARPN Journal of Engineering and Applied Sciences ©2006-2016 Asian Research Publishing Network (ARPN). All rights reserved. www.arpnjournals.com TWO-STEPS IMPLEMENTATION OF SIGMOID FUNCTION FOR ARTIFICIAL NEURAL NETWORK IN FIELD PROGRAMMABLE GATE ARRAY Syahrulanuar Ngah1, Rohani Abu Bakar1, Abdullah Embong1 and Saifudin Razali2 1Faculty of Computer Systems and Software Engineering, Universiti Malaysia Pahang, Malaysia 2Faculty of Electrical and Electronic Engineering, Universiti Malaysia Pahang, Malaysia [email protected] ABSTRACT The complex equation of sigmoid function is one of the most difficult problems encountered for implementing the artificial neural network (ANN) into a field programmable gate array (FPGA). To overcome this problem, the combination of second order nonlinear function (SONF) and the differential lookup table (dLUT) has been proposed in this paper. By using this two-steps approach, the output accuracy achieved is ten times better than that of using only SONF and two times better than that of using conventional lookup table (LUT). Hence, this method can be used in various applications that required the implementation of the ANN into FPGA. Keywords: sigmoid function, second order nonlinear function, differential lookup table, FPGA. INTRODUCTION namely Gaussian, logarithmic, hyperbolic tangent and The Artificial neural network (ANN) has become sigmoid function has been done by T. Tan et al. They a growing research focus over the last few decades. It have found that, the optimal configuration for their controller been used broadly in many fields to solve great variety of can be achieved by using the sigmoid function [12].In this problems such as speed estimation [1], image processing paper, sigmoid function is being used as an activation [2] and pattern recognition [3]. -
Information Flows of Diverse Autoencoders
entropy Article Information Flows of Diverse Autoencoders Sungyeop Lee 1,* and Junghyo Jo 2,3,* 1 Department of Physics and Astronomy, Seoul National University, Seoul 08826, Korea 2 Department of Physics Education and Center for Theoretical Physics and Artificial Intelligence Institute, Seoul National University, Seoul 08826, Korea 3 School of Computational Sciences, Korea Institute for Advanced Study, Seoul 02455, Korea * Correspondence: [email protected] (S.L.); [email protected] (J.J.) Abstract: Deep learning methods have had outstanding performances in various fields. A fundamen- tal query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visualized on the information plane of the mutual information among the input, hidden, and output layers. In this study, we examine how the information flows are shaped by the network parameters, such as depth, sparsity, weight constraints, and hidden representations. Here, we adopt autoencoders as models of deep learning, because (i) they have clear guidelines for their information flows, and (ii) they have various species, such as vanilla, sparse, tied, variational, and label autoen- coders. We measured their information flows using Rényi’s matrix-based a-order entropy functional. As learning progresses, they show a typical fitting phase where the amounts of input-to-hidden and hidden-to-output mutual information both increase. In the last stage of learning, however, some autoencoders show a simplifying phase, previously called the “compression phase”, where input-to-hidden mutual information diminishes. In particular, the sparsity regularization of hidden activities amplifies the simplifying phase. -
Compositional Distributional Semantics with Long Short Term Memory
Compositional Distributional Semantics with Long Short Term Memory Phong Le and Willem Zuidema Institute for Logic, Language and Computation University of Amsterdam, the Netherlands p.le,zuidema @uva.nl { } Abstract any continuous function (Cybenko, 1989) already make them an attractive choice. We are proposing an extension of the recur- In trying to answer the second question, the ad- sive neural network that makes use of a vari- vantages of approaches based on neural network ar- ant of the long short-term memory architec- ture. The extension allows information low chitectures, such as the recursive neural network in parse trees to be stored in a memory reg- (RNN) model (Socher et al., 2013b) and the con- ister (the ‘memory cell’) and used much later volutional neural network model (Kalchbrenner et higher up in the parse tree. This provides a so- al., 2014), are even clearer. Models in this paradigm lution to the vanishing gradient problem and can take advantage of general learning procedures allows the network to capture long range de- based on back-propagation, and with the rise of pendencies. Experimental results show that ‘deep learning’, of a variety of efficient algorithms our composition outperformed the traditional neural-network composition on the Stanford and tricks to further improve training. Sentiment Treebank. Since the first success of the RNN model (Socher et al., 2011b) in constituent parsing, two classes of extensions have been proposed. One class is to en- 1 Introduction hance its compositionality by using tensor product Moving from lexical to compositional semantics in (Socher et al., 2013b) or concatenating RNNs hor- vector-based semantics requires answers to two dif- izontally to make a deeper net (Irsoy and Cardie, ficult questions: (i) what is the nature of the com- 2014). -
Eis- a Family of Activation Functions Combining Exponential,Isru, and Softplus
EIS - A FAMILY OF ACTIVATION FUNCTIONS COMBINING EXPONENTIAL, ISRU, AND SOFTPLUS APREPRINT Koushik Biswas Sandeep Kumar Department of Computer Science Department of Computer Science, IIIT Delhi IIIT Delhi & New Delhi, India, 110020 Department of Mathematics, Shaheed Bhagat Singh College, [email protected] University of Delhi. New Delhi, India. [email protected], [email protected] Shilpak Banerjee Ashish Kumar Pandey Department of Mathematics Department of Mathematics IIIT Delhi IIIT Delhi New Delhi, India, 110020 New Delhi, India, 110020 [email protected] [email protected] ABSTRACT Activation functions play a pivotal role in the function learning using neural networks. The non- linearity in the learned function is achieved by repeated use of the activation function. Over the years, numerous activation functions have been proposed to improve accuracy in several tasks. Basic functions like ReLU, Sigmoid, Tanh, or Softplus have been favorite among the deep learning community because of their simplicity. In recent years, several novel activation functions arising from these basic functions have been proposed, which have improved accuracy in some challenging datasets. We propose a five hyper-parameters family of activation functions, namely EIS, defined as, x(ln(1 + ex))α : pβ + γx2 + δe−θx We show examples of activation functions from the EIS family which outperform widely used x ln(1+ex) activation functions on some well known datasets and models. For example, x+1:16e−x beats ReLU arXiv:2009.13501v2 [cs.LG] 12 Oct 2020 by 0.89% in DenseNet-169, 0.24% in Inception V3 in CIFAR100 dataset while 1.13% in Inception x ln(1+ex) V3, 0.13% in DenseNet-169, 0.94% in SimpleNet model in CIFAR10 dataset. -
Advancements in Image Classification Using Convolutional Neural Network
This paper has been accepted and presented on 2018 Fourth International Conference on Research in Computational Intelligence and Communication Networks. This is preprint version and original proceeding will be published in IEEE Xplore. Advancements in Image Classification using Convolutional Neural Network Farhana Sultana Abu Sufian Paramartha Dutta Department of Computer Science Department of Computer Science Department of CSS University of Gour Banga University of Gour Banga Visva-Bharati University West Bengal, India West Bengal, India West Bengal, India Email: [email protected] Email: sufi[email protected] Email: [email protected] Abstract—Convolutional Neural Network (CNN) is the state- LeCun et al. introduced the practical model of CNN [6] [7] of-the-art for image classification task. Here we have briefly and developed LeNet-5 [8]. Training by backpropagation [9] discussed different components of CNN. In this paper, We have algorithm helped LeNet-5 recognizing visual patterns from explained different CNN architectures for image classification. Through this paper, we have shown advancements in CNN from raw pixels directly without using any separate feature engi- LeNet-5 to latest SENet model. We have discussed the model neering mechanism. Also fewer connections and parameters description and training details of each model. We have also of CNN than conventional feedforward neural networks with drawn a comparison among those models. similar network size, made model training easier. But at that Keywords—AlexNet, Capsnet, Convolutional Neural Network, time in spite of several advantages, the performance of CNN Deep learning, DenseNet, Image classification, ResNet, SENet. in intricate problems such as classification of high-resolution image, was limited by the lack of large training data, lack of I. -
Global Sparse Momentum SGD for Pruning Very Deep Neural Networks
Global Sparse Momentum SGD for Pruning Very Deep Neural Networks Xiaohan Ding 1 Guiguang Ding 1 Xiangxin Zhou 2 Yuchen Guo 1, 3 Jungong Han 4 Ji Liu 5 1 Beijing National Research Center for Information Science and Technology (BNRist); School of Software, Tsinghua University, Beijing, China 2 Department of Electronic Engineering, Tsinghua University, Beijing, China 3 Department of Automation, Tsinghua University; Institute for Brain and Cognitive Sciences, Tsinghua University, Beijing, China 4 WMG Data Science, University of Warwick, Coventry, United Kingdom 5 Kwai Seattle AI Lab, Kwai FeDA Lab, Kwai AI platform [email protected] [email protected] [email protected] [email protected] [email protected] [email protected] Abstract Deep Neural Network (DNN) is powerful but computationally expensive and memory intensive, thus impeding its practical usage on resource-constrained front- end devices. DNN pruning is an approach for deep model compression, which aims at eliminating some parameters with tolerable performance degradation. In this paper, we propose a novel momentum-SGD-based optimization method to reduce the network complexity by on-the-fly pruning. Concretely, given a global compression ratio, we categorize all the parameters into two parts at each training iteration which are updated using different rules. In this way, we gradually zero out the redundant parameters, as we update them using only the ordinary weight decay but no gradients derived from the objective function. As a departure from prior methods that require heavy human works to tune the layer-wise sparsity ratios, prune by solving complicated non-differentiable problems or finetune the model after pruning, our method is characterized by 1) global compression that automatically finds the appropriate per-layer sparsity ratios; 2) end-to-end training; 3) no need for a time-consuming re-training process after pruning; and 4) superior capability to find better winning tickets which have won the initialization lottery.