An Advanced Pruning Method in the Architecture of Extreme Learning Machines Using L1-Regularization and Bootstrapping
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
ECS289: Scalable Machine Learning
ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Nov 3, 2015 Outline PageRank Semi-supervised Learning Label Propagation Manifold regularization PageRank Ranking Websites Text based ranking systems (a dominated approach in the early 90s) Compute the similarity between query and websites (documents) Keywords are a very limited way to express a complex information Need to rank websites by \popularity", \authority", . PageRank: Developed by Brin and Page (1999) Determine the authority and popularity by hyperlinks PageRank Main idea: estimate the ranking of websites by the link structure. Topology of Websites Transform the hyperlinks to a directed graph: The adjacency matrix A such that Aij = 1 if page j points to page i Transition Matrix Normalize the adjacency matrix so that the matrix is a stochastic matrix (each column sum up to 1) Pij : probability that arriving at page i from page j P: a stochastic matrix or a transition matrix Random walk: step 1 Random walk through the transition matrix Start from [1; 0; 0; 0] (can use any initialization) Random walk: step 1 Random walk through the transition matrix xt+1 = Pxt Random walk: step 2 Random walk through the transition matrix Random walk: step 2 Random walk through the transition matrix xt+2 = Pxt+1 PageRank (convergence) P Start from an initial vector x with i xi = 1 (the initial distribution) For t = 1; 2;::: xt+1 = Pxt Each xt is a probability distribution (sums up to 1) Converges to a stationary distribution π such that π = Pπ if P satisfies the following two conditions: t 1 P is irreducible: for all i; j, there exists some t such that (P )i;j > 0 t 2 P is aperiodic: for all i; j, we have gcdft :(P )i;j > 0g = 1 π is the unique right eigenvector of P with eigenvalue 1. -
Semi-Supervised Classification Using Local and Global Regularization
Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (2008) Semi- supervised Classification Using Local and Global Regularization Fei Wang1, Tao Li2, Gang Wang3, Changshui Zhang1 1Department of Automation, Tsinghua University, Beijing, China 2School of Computing and Information Sciences, Florida International University, Miami, FL, USA 3Microsoft China Research, Beijing, China Abstract works concentrate on the derivation of different forms of In this paper, we propose a semi-supervised learning (SSL) smoothness regularizers, such as the ones using combinato- algorithm based on local and global regularization. In the lo- rial graph Laplacian (Zhu et al., 2003)(Belkin et al., 2006), cal regularization part, our algorithm constructs a regularized normalized graph Laplacian (Zhou et al., 2004), exponen- classifier for each data point using its neighborhood, while tial/iterative graph Laplacian (Belkin et al., 2004), local lin- the global regularization part adopts a Laplacian regularizer ear regularization (Wang & Zhang, 2006)and local learning to smooth the data labels predicted by those local classifiers. regularization (Wu & Sch¨olkopf, 2007), but rarely touch the We show that some existing SSL algorithms can be derived problem of how to derive a more efficient loss function. from our framework. Finally we present some experimental In this paper, we argue that rather than applying a global results to show the effectiveness of our method. loss function which is based on the construction of a global predictor using the whole data set, it would be more desir- Introduction able to measure such loss locally by building some local pre- Semi-supervised learning (SSL), which aims at learning dictors for different regions of the input data space. -
Getting Started with Machine Learning
Getting Started with Machine Learning CSC131 The Beauty & Joy of Computing Cornell College 600 First Street SW Mount Vernon, Iowa 52314 September 2018 ii Contents 1 Applications: where machine learning is helping1 1.1 Sheldon Branch............................1 1.2 Bram Dedrick.............................4 1.3 Tony Ferenzi.............................5 1.3.1 Benefits of Machine Learning................5 1.4 William Golden............................7 1.4.1 Humans: The Teachers of Technology...........7 1.5 Yuan Hong..............................9 1.6 Easton Jensen............................. 11 1.7 Rodrigo Martinez........................... 13 1.7.1 Machine Learning in Medicine............... 13 1.8 Matt Morrical............................. 15 1.9 Ella Nelson.............................. 16 1.10 Koichi Okazaki............................ 17 1.11 Jakob Orel.............................. 19 1.12 Marcellus Parks............................ 20 1.13 Lydia Sanchez............................. 22 1.14 Tiff Serra-Pichardo.......................... 24 1.15 Austin Stala.............................. 25 1.16 Nicole Trenholm........................... 26 1.17 Maddy Weaver............................ 28 1.18 Peter Weber.............................. 29 iii iv CONTENTS 2 Recommendations: How to learn more about machine learning 31 2.1 Sheldon Branch............................ 31 2.1.1 Course 1: Machine Learning................. 31 2.1.2 Course 2: Robotics: Vision Intelligence and Machine Learn- ing............................... 33 2.1.3 Course -
Linear Manifold Regularization for Large Scale Semi-Supervised Learning
Linear Manifold Regularization for Large Scale Semi-supervised Learning Vikas Sindhwani [email protected] Partha Niyogi [email protected] Mikhail Belkin [email protected] Department of Computer Science, University of Chicago, Hyde Park, Chicago, IL 60637 Sathiya Keerthi [email protected] Yahoo! Research Labs, 210 S Delacey Avenue, Pasadena, CA 91105 Abstract tion (Belkin, Niyogi & Sindhwani, 2004; Sindhwani, The enormous wealth of unlabeled data in Niyogi and Belkin, 2005). In this framework, unla- many applications of machine learning is be- beled data is incorporated within a geometrically mo- ginning to pose challenges to the designers tivated regularizer. We specialize this framework for of semi-supervised learning methods. We linear classification, and focus on the problem of deal- are interested in developing linear classifica- ing with large amounts of data. tion algorithms to efficiently learn from mas- This paper is organized as follows: In section 2, we sive partially labeled datasets. In this paper, setup the problem of linear semi-supervised learning we propose Linear Laplacian Support Vector and discuss some prior work. The general Manifold Machines and Linear Laplacian Regularized Regularization framework is reviewed in section 3. In Least Squares as promising solutions to this section 4, we discuss our methods. Section 5 outlines problem. some research in progress. 1. Introduction 2. Linear Semi-supervised Learning The problem of linear semi-supervised clasification is A single web-crawl by search engines like Yahoo and setup as follows: We are given l labeled examples Google indexes billions of webpages. Only a very l l+u fxi; yigi=1 and u unlabeled examples fxigi=l+1, where small fraction of these web-pages can be hand-labeled d and assembled into topic directories. -
Semi-Supervised Deep Learning Using Improved Unsupervised Discriminant Projection?
Semi-Supervised Deep Learning Using Improved Unsupervised Discriminant Projection? Xiao Han∗, Zihao Wang∗, Enmei Tu, Gunnam Suryanarayana, and Jie Yang School of Electronics, Information and Electrical Engineering, Shanghai Jiao Tong University, China [email protected], [email protected], [email protected] Abstract. Deep learning demands a huge amount of well-labeled data to train the network parameters. How to use the least amount of labeled data to obtain the desired classification accuracy is of great practical significance, because for many real-world applications (such as medical diagnosis), it is difficult to obtain so many labeled samples. In this paper, modify the unsupervised discriminant projection algorithm from dimen- sion reduction and apply it as a regularization term to propose a new semi-supervised deep learning algorithm, which is able to utilize both the local and nonlocal distribution of abundant unlabeled samples to improve classification performance. Experiments show that given dozens of labeled samples, the proposed algorithm can train a deep network to attain satisfactory classification results. Keywords: Manifold Regularization · Semi-supervised learning · Deep learning. 1 Introduction In reality, one of the main difficulties faced by many machine learning tasks is manually tagging large amounts of data. This is especially prominent for deep learning, which usually demands a huge number of well-labeled samples. There- fore, how to use the least amount of labeled data to train a deep network has become an important topic in the area. To overcome this problem, researchers proposed that the use of a large number of unlabeled data can extract the topol- ogy of the overall data's distribution. -
Hyper-Parameter Optimization for Manifold Regularization Learning
Cassiano Ot´avio Becker Hyper-parameter Optimization for Manifold Regularization Learning Otimiza¸cao~ de Hiperparametros^ para Aprendizado do Computador por Regulariza¸cao~ em Variedades Campinas 2013 i ii Universidade Estadual de Campinas Faculdade de Engenharia Eletrica´ e de Computa¸cao~ Cassiano Ot´avio Becker Hyper-parameter Optimization for Manifold Regularization Learning Otimiza¸cao~ de Hiperparametros^ para Aprendizado do Computador por Regulariza¸cao~ em Variedades Master dissertation presented to the School of Electrical and Computer Engineering in partial fulfillment of the requirements for the degree of Master in Electrical Engineering. Concentration area: Automation. Disserta¸c~ao de Mestrado apresentada ao Programa de P´os- Gradua¸c~ao em Engenharia El´etrica da Faculdade de Engenharia El´etrica e de Computa¸c~ao da Universidade Estadual de Campinas como parte dos requisitos exigidos para a obten¸c~ao do t´ıtulo de Mestre em Engenharia El´etrica. Area´ de concentra¸c~ao: Automa¸c~ao. Orientador (Advisor): Prof. Dr. Paulo Augusto Valente Ferreira Este exemplar corresponde `a vers~ao final da disserta¸c~aodefendida pelo aluno Cassiano Ot´avio Becker, e orientada pelo Prof. Dr. Paulo Augusto Valente Ferreira. Campinas 2013 iii Ficha catalográfica Universidade Estadual de Campinas Biblioteca da Área de Engenharia e Arquitetura Rose Meire da Silva - CRB 8/5974 Becker, Cassiano Otávio, 1977- B388h BecHyper-parameter optimization for manifold regularization learning / Cassiano Otávio Becker. – Campinas, SP : [s.n.], 2013. BecOrientador: Paulo Augusto Valente Ferreira. BecDissertação (mestrado) – Universidade Estadual de Campinas, Faculdade de Engenharia Elétrica e de Computação. Bec1. Aprendizado do computador. 2. Aprendizado semi-supervisionado. 3. Otimização matemática. 4. -
Zero Shot Learning Via Multi-Scale Manifold Regularization
Zero Shot Learning via Multi-Scale Manifold Regularization Shay Deutsch 1,2 Soheil Kolouri 3 Kyungnam Kim 3 Yuri Owechko 3 Stefano Soatto 1 1 UCLA Vision Lab, University of California, Los Angeles, CA 90095 2 Department of Mathematics, University of California Los Angeles 3 HRL Laboratories, LLC, Malibu, CA {shaydeu@math., soatto@cs.}ucla.edu {skolouri, kkim, yowechko,}@hrl.com Abstract performed in a transductive setting, using all unlabeled data where the classification and learning process on the transfer We address zero-shot learning using a new manifold data is entirely unsupervised. alignment framework based on a localized multi-scale While our suggested approach is similar in scope to other transform on graphs. Our inference approach includes a ”visual-semantic alignment” methods (see [5, 10] and ref- smoothness criterion for a function mapping nodes on a erences therein) it is, to the best of our knowledge, the first graph (visual representation) onto a linear space (semantic to use multi-scale localized representations that respect the representation), which we optimize using multi-scale graph global (non-flat) geometry of the data space and the fine- wavelets. The robustness of the ensuing scheme allows us scale structure of the local semantic representation. More- to operate with automatically generated semantic annota- over, learning the relationship between visual features and tions, resulting in an algorithm that is entirely free of man- semantic attributes is unified into a single process, whereas ual supervision, and yet improves the state-of-the-art as in most ZSL approaches it is divided into a number of inde- measured on benchmark datasets. -
Deep Hashing Using an Extreme Learning Machine with Convolutional Networks
i \1-Zeng" | 2018/2/2 | 23:57 | page 133 | #1 i i i Communications in Information and Systems Volume 17, Number 3, 133{146, 2017 Deep hashing using an extreme learning machine with convolutional networks Zhiyong Zeng∗, Shiqi Dai, Yunsong Li, Dunyu Chen In this paper, we present a deep hashing approach for large scale image search. It is different from most existing deep hash learn- ing methods which use convolutional neural networks (CNN) to execute feature extraction to learn binary codes. These methods could achieve excellent results, but they depend on an extreme huge and complex networks. We combine an extreme learning ma- chine (ELM) with convolutional neural networks to speed up the training of deep learning methods. In contrast to existing deep hashing approaches, our method leads to faster and more accurate feature learning. Meanwhile, it improves the generalization ability of deep hashing. Experiments on large scale image datasets demon- strate that the proposed approach can achieve better results with state-of-the-art methods with much less complexity. 1. Introduction With the growing of multimedia data, fast and effective search technique has become a hot research topic. Among existing search techniques, hashing is one of the most important retrieval techniques due to its fast query speed and low memory cost. Hashing methods can be divided into two categories: data-independent [1{3], data-dependent [4{8]. For the first category, hash functions random generated are first used to map data into feature space and then binarization is carried out. Representative methods of this category are locality sensitive hashing (LSH) [1] and its variants [2, 3]. -
Semi-Supervised Deep Metric Learning Networks for Classification of Polarimetric SAR Data
remote sensing Letter Semi-Supervised Deep Metric Learning Networks for Classification of Polarimetric SAR Data Hongying Liu , Ruyi Luo, Fanhua Shang * , Xuechun Meng, Shuiping Gou and Biao Hou Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an 710071, China; [email protected] (H.L.); [email protected] (R.L.); [email protected] (X.M.); [email protected] (S.G.); [email protected] (B.H.) * Corresponding authors: [email protected] Received: 7 April 2020; Accepted: 14 May 2020; Published: 17 May 2020 Abstract: Recently, classification methods based on deep learning have attained sound results for the classification of Polarimetric synthetic aperture radar (PolSAR) data. However, they generally require a great deal of labeled data to train their models, which limits their potential real-world applications. This paper proposes a novel semi-supervised deep metric learning network (SSDMLN) for feature learning and classification of PolSAR data. Inspired by distance metric learning, we construct a network, which transforms the linear mapping of metric learning into the non-linear projection in the layer-by-layer learning. With the prior knowledge of the sample categories, the network also learns a distance metric under which all pairs of similarly labeled samples are closer and dissimilar samples have larger relative distances. Moreover, we introduce a new manifold regularization to reduce the distance between neighboring samples since they are more likely to be homogeneous. The categorizing is achieved by using a simple classifier. Several experiments on both synthetic and real-world PolSAR data from different sensors are conducted and they demonstrate the effectiveness of SSDMLN with limited labeled samples, and SSDMLN is superior to state-of-the-art methods. -
Convolutional Neural Network Based on Extreme Learning Machine for Maritime Ships Recognition in Infrared Images
sensors Article Convolutional Neural Network Based on Extreme Learning Machine for Maritime Ships Recognition in Infrared Images Atmane Khellal 1,*, Hongbin Ma 1,2,* and Qing Fei 1,2 1 School of Automation, Beijing Institute of Technology, Beijing 100081, China; [email protected] 2 State Key Laboratory of Intelligent Control and Decision of Complex Systems, Beijing Institute of Technology, Beijing 100081, China * Correspondence: [email protected] (A.K.); [email protected] (H.M.); Tel.: +86-132-6107-8800 (A.K.); +86-152-1065-8048 (H.M.) Received: 1 April 2018; Accepted: 7 May 2018; Published: 9 May 2018 Abstract: The success of Deep Learning models, notably convolutional neural networks (CNNs), makes them the favorable solution for object recognition systems in both visible and infrared domains. However, the lack of training data in the case of maritime ships research leads to poor performance due to the problem of overfitting. In addition, the back-propagation algorithm used to train CNN is very slow and requires tuning many hyperparameters. To overcome these weaknesses, we introduce a new approach fully based on Extreme Learning Machine (ELM) to learn useful CNN features and perform a fast and accurate classification, which is suitable for infrared-based recognition systems. The proposed approach combines an ELM based learning algorithm to train CNN for discriminative features extraction and an ELM based ensemble for classification. The experimental results on VAIS dataset, which is the largest dataset of maritime ships, confirm that the proposed approach outperforms the state-of-the-art models in term of generalization performance and training speed. -
A Real-Time Sequential Deep Extreme Learning Machine Cybersecurity Intrusion Detection System
Computers, Materials & Continua Tech Science Press DOI:10.32604/cmc.2020.013910 Article A Real-Time Sequential Deep Extreme Learning Machine Cybersecurity Intrusion Detection System Amir Haider1, Muhammad Adnan Khan2, Abdur Rehman3, Muhib Ur Rahman4 and Hyung Seok Kim1,* 1Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, 05006, Korea 2Department of Computer Science, Lahore Garrison University, Lahore, 54000, Pakistan 3School of Computer Science, National College of Business Administration & Economics, Lahore, 54000, Pakistan 4Department of Electrical Engineering, Polytechnique Montreal, Montreal, QC H3T 1J4, Canada ÃCorresponding Author: Hyung Seok Kim. Email: [email protected] Received: 26 August 2020; Accepted: 28 September 2020 Abstract: In recent years, cybersecurity has attracted significant interest due to the rapid growth of the Internet of Things (IoT) and the widespread development of computer infrastructure and systems. It is thus becoming particularly necessary to identify cyber-attacks or irregularities in the system and develop an efficient intru- sion detection framework that is integral to security. Researchers have worked on developing intrusion detection models that depend on machine learning (ML) methods to address these security problems. An intelligent intrusion detection device powered by data can exploit artificial intelligence (AI), and especially ML, techniques. Accordingly, we propose in this article an intrusion detection model based on a Real-Time Sequential Deep Extreme Learning Machine Cyber- security Intrusion Detection System (RTS-DELM-CSIDS) security model. The proposed model initially determines the rating of security aspects contributing to their significance and then develops a comprehensive intrusion detection frame- work focused on the essential characteristics. Furthermore, we investigated the feasibility of our proposed RTS-DELM-CSIDS framework by performing dataset evaluations and calculating accuracy parameters to validate. -
Manifold Regularization and Semi-Supervised Learning: Some Theoretical Analyses
Manifold Regularization and Semi-supervised Learning: Some Theoretical Analyses Partha Niyogi∗ January 22, 2008 Abstract Manifold regularization (Belkin, Niyogi, Sindhwani, 2004)isage- ometrically motivated framework for machine learning within which several semi-supervised algorithms have been constructed. Here we try to provide some theoretical understanding of this approach. Our main result is to expose the natural structure of a class of problems on which manifold regularization methods are helpful. We show that for such problems, no supervised learner can learn effectively. On the other hand, a manifold based learner (that knows the manifold or “learns” it from unlabeled examples) can learn with relatively few labeled examples. Our analysis follows a minimax style with an em- phasis on finite sample results (in terms of n: the number of labeled examples). These results allow us to properly interpret manifold reg- ularization and its potential use in semi-supervised learning. 1 Introduction A learning problem is specified by a probability distribution p on X × Y according to which labelled examples zi = (xi, yi) pairs are drawn and presented to a learning algorithm (estimation procedure). We are interested in an understanding of the case in which X = IRD, ∗Departments of Computer Science, Statistics, The University of Chicago 1 Y ⊂ IR but pX (the marginal distribution of p on X) is supported on some submanifold M ⊂ X. In particular, we are interested in understanding how knowledge of this submanifold may potentially help a learning algorithm1. To this end, we will consider two kinds of learning algorithms: 1. Algorithms that have no knowledge of the submanifold M but learn from (xi, yi) pairs in a purely supervised way.