Automated Feature Engineering for Predictive Modeling
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
10 Oriented Principal Component Analysis for Feature Extraction
10 Oriented Principal Component Analysis for Feature Extraction Abstract- Principal Components Analysis (PCA) is a popular unsupervised learning technique that is often used in pattern recognition for feature extraction. Some variants have been proposed for supervising PCA to include class information, which are usually based on heuristics. A more principled approach is presented in this paper. Suppose that we have a pattern recogniser composed of a linear network for feature extraction and a classifier. Instead of designing each module separately, we perform global training where both learning systems cooperate in order to find those components (Oriented PCs) useful for classification. Experimental results, using artificial and real data, show the potential of the proposed learning system for classification. Since the linear neural network finds the optimal directions to better discriminate classes, the global-trained pattern recogniser needs less Oriented PCs (OPCs) than PCs so the classifier’s complexity (e.g. capacity) is lower. In this way, the global system benefits from a better generalisation performance. Index Terms- Oriented Principal Components Analysis, Principal Components Neural Networks, Co-operative Learning, Learning to Learn Algorithms, Feature Extraction, Online gradient descent, Pattern Recognition, Compression. Abbreviations- PCA- Principal Components Analysis, OPC- Oriented Principal Components, PCNN- Principal Components Neural Networks. 1. Introduction In classification, observations belong to different classes. Based on some prior knowledge about the problem and on the training set, a pattern recogniser is constructed for assigning future observations to one of the existing classes. The typical method of building this system consists in dividing it into two main blocks that are usually trained separately: the feature extractor and the classifier. -
Incorporating Automated Feature Engineering Routines Into Automated Machine Learning Pipelines by Wesley Runnels
Incorporating Automated Feature Engineering Routines into Automated Machine Learning Pipelines by Wesley Runnels Submitted to the Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree of Master of Engineering in Electrical Engineering and Computer Science at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY May 2020 c Massachusetts Institute of Technology 2020. All rights reserved. Author . Wesley Runnels Department of Electrical Engineering and Computer Science May 12th, 2020 Certified by . Tim Kraska Department of Electrical Engineering and Computer Science Thesis Supervisor Accepted by . Katrina LaCurts Chair, Master of Engineering Thesis Committee 2 Incorporating Automated Feature Engineering Routines into Automated Machine Learning Pipelines by Wesley Runnels Submitted to the Department of Electrical Engineering and Computer Science on May 12, 2020, in partial fulfillment of the requirements for the degree of Master of Engineering in Electrical Engineering and Computer Science Abstract Automating the construction of consistently high-performing machine learning pipelines has remained difficult for researchers, especially given the domain knowl- edge and expertise often necessary for achieving optimal performance on a given dataset. In particular, the task of feature engineering, a key step in achieving high performance for machine learning tasks, is still mostly performed manually by ex- perienced data scientists. In this thesis, building upon the results of prior work in this domain, we present a tool, rl feature eng, which automatically generates promising features for an arbitrary dataset. In particular, this tool is specifically adapted to the requirements of augmenting a more general auto-ML framework. We discuss the performance of this tool in a series of experiments highlighting the various options available for use, and finally discuss its performance when used in conjunction with Alpine Meadow, a general auto-ML package. -
Feature Selection/Extraction
Feature Selection/Extraction Dimensionality Reduction Feature Selection/Extraction • Solution to a number of problems in Pattern Recognition can be achieved by choosing a better feature space. • Problems and Solutions: – Curse of Dimensionality: • #examples needed to train classifier function grows exponentially with #dimensions. • Overfitting and Generalization performance – What features best characterize class? • What words best characterize a document class • Subregions characterize protein function? – What features critical for performance? – Subregions characterize protein function? – Inefficiency • Reduced complexity and run-time – Can’t Visualize • Allows ‘intuiting’ the nature of the problem solution. Curse of Dimensionality Same Number of examples Fill more of the available space When the dimensionality is low Selection vs. Extraction • Two general approaches for dimensionality reduction – Feature extraction: Transforming the existing features into a lower dimensional space – Feature selection: Selecting a subset of the existing features without a transformation • Feature extraction – PCA – LDA (Fisher’s) – Nonlinear PCA (kernel, other varieties – 1st layer of many networks Feature selection ( Feature Subset Selection ) Although FS is a special case of feature extraction, in practice quite different – FSS searches for a subset that minimizes some cost function (e.g. test error) – FSS has a unique set of methodologies Feature Subset Selection Definition Given a feature set x={xi | i=1…N} find a subset xM ={xi1, xi2, …, xiM}, with M<N, that optimizes an objective function J(Y), e.g. P(correct classification) Why Feature Selection? • Why not use the more general feature extraction methods? Feature Selection is necessary in a number of situations • Features may be expensive to obtain • Want to extract meaningful rules from your classifier • When you transform or project, measurement units (length, weight, etc.) are lost • Features may not be numeric (e.g. -
Feature Extraction (PCA & LDA)
Feature Extraction (PCA & LDA) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline What is feature extraction? Feature extraction algorithms Linear Methods Unsupervised: Principal Component Analysis (PCA) Also known as Karhonen-Loeve (KL) transform Supervised: Linear Discriminant Analysis (LDA) Also known as Fisher’s Discriminant Analysis (FDA) 2 Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given feature set Feature extraction (e.g., PCA, LDA) A linear or non-linear transform on the original feature space ⋮ ⋮ → ⋮ → ⋮ = ⋮ Feature Selection Feature ( <) Extraction 3 Feature Extraction Mapping of the original data onto a lower-dimensional space Criterion for feature extraction can be different based on problem settings Unsupervised task: minimize the information loss (reconstruction error) Supervised task: maximize the class discrimination on the projected space In the previous lecture, we talked about feature selection: Feature selection can be considered as a special form of feature extraction (only a subset of the original features are used). Example: 0100 X N 4 XX' 0010 X' N 2 Second and thirth features are selected 4 Feature Extraction Unsupervised feature extraction: () () A mapping : ℝ →ℝ ⋯ Or = ⋮⋱⋮ Feature Extraction only the transformed data () () () () ⋯ ′ ⋯′ = ⋮⋱⋮ () () ′ ⋯′ Supervised feature extraction: () () A mapping : ℝ →ℝ ⋯ Or = ⋮⋱⋮ Feature Extraction only the transformed -
Feature Extraction for Image Selection Using Machine Learning
Master of Science Thesis in Electrical Engineering Department of Electrical Engineering, Linköping University, 2017 Feature extraction for image selection using machine learning Matilda Lorentzon Master of Science Thesis in Electrical Engineering Feature extraction for image selection using machine learning Matilda Lorentzon LiTH-ISY-EX--17/5097--SE Supervisor: Marcus Wallenberg ISY, Linköping University Tina Erlandsson Saab Aeronautics Examiner: Lasse Alfredsson ISY, Linköping University Computer Vision Laboratory Department of Electrical Engineering Linköping University SE-581 83 Linköping, Sweden Copyright © 2017 Matilda Lorentzon Abstract During flights with manned or unmanned aircraft, continuous recording can result in a very high number of images to analyze and evaluate. To simplify image analysis and to minimize data link usage, appropriate images should be suggested for transfer and further analysis. This thesis investigates features used for selection of images worthy of further analysis using machine learning. The selection is done based on the criteria of having good quality, salient content and being unique compared to the other selected images. The investigation is approached by implementing two binary classifications, one regard- ing content and one regarding quality. The classifications are made using support vector machines. For each of the classifications three feature extraction methods are performed and the results are compared against each other. The feature extraction methods used are histograms of oriented gradients, features from the discrete cosine transform domain and features extracted from a pre-trained convolutional neural network. The images classified as both good and salient are then clustered based on similarity measures retrieved using color coherence vectors. One image from each cluster is retrieved and those are the result- ing images from the image selection. -
Learning from Few Subjects with Large Amounts of Voice Monitoring Data
Learning from Few Subjects with Large Amounts of Voice Monitoring Data Jose Javier Gonzalez Ortiz John V. Guttag Robert E. Hillman Daryush D. Mehta Jarrad H. Van Stan Marzyeh Ghassemi Challenges of Many Medical Time Series • Few subjects and large amounts of data → Overfitting to subjects • No obvious mapping from signal to features → Feature engineering is labor intensive • Subject-level labels → In many cases, no good way of getting sample specific annotations 1 Jose Javier Gonzalez Ortiz Challenges of Many Medical Time Series • Few subjects and large amounts of data → Overfitting to subjects Unsupervised feature • No obvious mapping from signal to features extraction → Feature engineering is labor intensive • Subject-level labels Multiple → In many cases, no good way of getting sample Instance specific annotations Learning 2 Jose Javier Gonzalez Ortiz Learning Features • Segment signal into windows • Compute time-frequency representation • Unsupervised feature extraction Conv + BatchNorm + ReLU 128 x 64 Pooling Dense 128 x 64 Upsampling Sigmoid 3 Jose Javier Gonzalez Ortiz Classification Using Multiple Instance Learning • Logistic regression on learned features with subject labels RawRaw LogisticLogistic WaveformWaveform SpectrogramSpectrogram EncoderEncoder RegressionRegression PredictionPrediction •• %% Positive Positive Ours PerPer Window Window PerPer Subject Subject • Aggregate prediction using % positive windows per subject 4 Jose Javier Gonzalez Ortiz Application: Voice Monitoring Data • Voice disorders affect 7% of the US population • Data collected through neck placed accelerometer 1 week = ~4 billion samples ~100 5 Jose Javier Gonzalez Ortiz Results Previous work relied on expert designed features[1] AUC Accuracy Comparable performance Train 0.70 ± 0.05 0.71 ± 0.04 Expert LR without Test 0.68 ± 0.05 0.69 ± 0.04 task-specific feature Train 0.73 ± 0.06 0.72 ± 0.04 engineering! Ours Test 0.69 ± 0.07 0.70 ± 0.05 [1] Marzyeh Ghassemi et al. -
Time Series Feature Extraction for Industrial Big Data (Iiot) Applications
Time Series Feature Extraction for Industrial Big Data (IIoT) Applications [1] Term paper for the subject „Current Topics of Data Engineering” In the course of studies MSc. Data Engineering at the Jacobs University Bremen Markus Sun Born on 07.11.1994 in Achim Student/ Registration number: 30003082 1. Reviewer: Prof. Dr. Adalbert F.X. Wilhelm 2. Reviewer: Prof. Dr. Stefan Kettemann Time Series Feature Extraction for Industrial Big Data (IIoT) Applications __________________________________________________________________________ Table of Content Abstract ...................................................................................................................... 4 1 Introduction ...................................................................................................... 4 2 Main Topic - Terminology ................................................................................. 5 2.1 Sensor Data .............................................................................................. 5 2.2 Feature Selection and Feature Extraction...................................................... 5 2.2.1 Variance Threshold.............................................................................................................. 7 2.2.2 Correlation Threshold .......................................................................................................... 7 2.2.3 Principal Component Analysis (PCA) .................................................................................. 7 2.2.4 Linear Discriminant Analysis -
Machine Learning Feature Extraction Based on Binary Pixel Quantification Using Low-Resolution Images for Application of Unmanned Ground Vehicles in Apple Orchards
agronomy Article Machine Learning Feature Extraction Based on Binary Pixel Quantification Using Low-Resolution Images for Application of Unmanned Ground Vehicles in Apple Orchards Hong-Kun Lyu 1,*, Sanghun Yun 1 and Byeongdae Choi 1,2 1 Division of Electronics and Information System, ICT Research Institute, Daegu Gyeongbuk Institute of Science and Technology (DGIST), Daegu 42988, Korea; [email protected] (S.Y.); [email protected] (B.C.) 2 Department of Interdisciplinary Engineering, Daegu Gyeongbuk Institute of Science and Technology (DGIST), Daegu 42988, Korea * Correspondence: [email protected]; Tel.: +82-53-785-3550 Received: 20 November 2020; Accepted: 4 December 2020; Published: 8 December 2020 Abstract: Deep learning and machine learning (ML) technologies have been implemented in various applications, and various agriculture technologies are being developed based on image-based object recognition technology. We propose an orchard environment free space recognition technology suitable for developing small-scale agricultural unmanned ground vehicle (UGV) autonomous mobile equipment using a low-cost lightweight processor. We designed an algorithm to minimize the amount of input data to be processed by the ML algorithm through low-resolution grayscale images and image binarization. In addition, we propose an ML feature extraction method based on binary pixel quantification that can be applied to an ML classifier to detect free space for autonomous movement of UGVs from binary images. Here, the ML feature is extracted by detecting the local-lowest points in segments of a binarized image and by defining 33 variables, including local-lowest points, to detect the bottom of a tree trunk. We trained six ML models to select a suitable ML model for trunk bottom detection among various ML models, and we analyzed and compared the performance of the trained models. -
Expert Feature-Engineering Vs. Deep Neural Networks: Which Is Better for Sensor-Free Affect Detection?
Expert Feature-Engineering vs. Deep Neural Networks: Which is Better for Sensor-Free Affect Detection? Yang Jiang1, Nigel Bosch2, Ryan S. Baker3, Luc Paquette2, Jaclyn Ocumpaugh3, Juliana Ma. Alexandra L. Andres3, Allison L. Moore4, Gautam Biswas4 1 Teachers College, Columbia University, New York, NY, United States [email protected] 2 University of Illinois at Urbana-Champaign, Champaign, IL, United States {pnb, lpaq}@illinois.edu 3 University of Pennsylvania, Philadelphia, PA, United States {rybaker, ojaclyn}@upenn.edu [email protected] 4 Vanderbilt University, Nashville, TN, United States {allison.l.moore, gautam.biswas}@vanderbilt.edu Abstract. The past few years have seen a surge of interest in deep neural net- works. The wide application of deep learning in other domains such as image classification has driven considerable recent interest and efforts in applying these methods in educational domains. However, there is still limited research compar- ing the predictive power of the deep learning approach with the traditional feature engineering approach for common student modeling problems such as sensor- free affect detection. This paper aims to address this gap by presenting a thorough comparison of several deep neural network approaches with a traditional feature engineering approach in the context of affect and behavior modeling. We built detectors of student affective states and behaviors as middle school students learned science in an open-ended learning environment called Betty’s Brain, us- ing both approaches. Overall, we observed a tradeoff where the feature engineer- ing models were better when considering a single optimized threshold (for inter- vention), whereas the deep learning models were better when taking model con- fidence fully into account (for discovery with models analyses). -
Towards Reproducible Meta-Feature Extraction
Journal of Machine Learning Research 21 (2020) 1-5 Submitted 5/19; Revised 12/19; Published 1/20 MFE: Towards reproducible meta-feature extraction Edesio Alcobaça [email protected] Felipe Siqueira [email protected] Adriano Rivolli [email protected] Luís P. F. Garcia [email protected] Jefferson T. Oliva [email protected] André C. P. L. F. de Carvalho [email protected] Institute of Mathematical and Computer Sciences University of São Paulo Av. Trabalhador São-carlense, 400, São Carlos, São Paulo 13560-970, Brazil Editor: Alexandre Gramfort Abstract Automated recommendation of machine learning algorithms is receiving a large deal of attention, not only because they can recommend the most suitable algorithms for a new task, but also because they can support efficient hyper-parameter tuning, leading to better machine learning solutions. The automated recommendation can be implemented using meta-learning, learning from previous learning experiences, to create a meta-model able to associate a data set to the predictive performance of machine learning algorithms. Al- though a large number of publications report the use of meta-learning, reproduction and comparison of meta-learning experiments is a difficult task. The literature lacks extensive and comprehensive public tools that enable the reproducible investigation of the differ- ent meta-learning approaches. An alternative to deal with this difficulty is to develop a meta-feature extractor package with the main characterization measures, following uniform guidelines that facilitate the use and inclusion of new meta-features. In this paper, we pro- pose two Meta-Feature Extractor (MFE) packages, written in both Python and R, to fill this lack. -
Generalized Feature Extraction for Structural Pattern Recognition in Time-Series Data Robert T
Generalized Feature Extraction for Structural Pattern Recognition in Time-Series Data Robert T. Olszewski February 2001 CMU-CS-01-108 School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy Thesis Committee: Roy Maxion, co-chair Dan Siewiorek, co-chair Christos Faloutsos David Banks, DOT Copyright c 2001 Robert T. Olszewski This research was sponsored by the Defense Advanced Project Research Agency (DARPA) and the Air Force Research Laboratory (AFRL) under grant #F30602-96-1-0349, the National Science Foundation (NSF) under grant #IRI-9224544, and the Air Force Laboratory Graduate Fellowship Program sponsored by the Air Force Office of Scientific Research (AFOSR) and conducted by the Southeastern Center for Electrical Engineering Education (SCEEE). The views and conclusions contained in this document are those of the author and should not be interpreted as representing the official policies, either expressed or implied, of DARPA, AFRL, NSF, AFOSR, SCEEE, or the U.S. government. Keywords: Structural pattern recognition, classification, feature extraction, time series, Fourier transformation, wavelets, semiconductor fabrication, electrocardiography. Abstract Pattern recognition encompasses two fundamental tasks: description and classification. Given an object to analyze, a pattern recognition system first generates a description of it (i.e., the pat- tern) and then classifies the object based on that description (i.e., the recognition). Two general approaches for implementing pattern recognition systems, statistical and structural, employ differ- ent techniques for description and classification. Statistical approaches to pattern recognition use decision-theoretic concepts to discriminate among objects belonging to different groups based upon their quantitative features. -
Feature Extraction Using Dimensionality Reduction Techniques: Capturing the Human Perspective
Wright State University CORE Scholar Browse all Theses and Dissertations Theses and Dissertations 2015 Feature Extraction using Dimensionality Reduction Techniques: Capturing the Human Perspective Ashley B. Coleman Wright State University Follow this and additional works at: https://corescholar.libraries.wright.edu/etd_all Part of the Computer Engineering Commons, and the Computer Sciences Commons Repository Citation Coleman, Ashley B., "Feature Extraction using Dimensionality Reduction Techniques: Capturing the Human Perspective" (2015). Browse all Theses and Dissertations. 1630. https://corescholar.libraries.wright.edu/etd_all/1630 This Thesis is brought to you for free and open access by the Theses and Dissertations at CORE Scholar. It has been accepted for inclusion in Browse all Theses and Dissertations by an authorized administrator of CORE Scholar. For more information, please contact [email protected]. FEATURE EXTRACTION USING DIMENSIONALITY REDUCITON TECHNIQUES: CAPTURING THE HUMAN PERSPECTIVE A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science By Ashley Coleman B.S., Wright State University, 2010 2015 Wright State University WRIGHT STATE UNIVERSITY GRADUATE SCHOOL December 11, 2015 I HEREBY RECOMMEND THAT THE THESIS PREPARED UNDER MY SUPERVISION BY Ashley Coleman ENTI TLED Feature Extraction using Dimensionality Reduction Techniques: Capturing the Human Perspective BE ACCEPTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Master of Science. Pascal Hitzler, Ph.D. Thesis Advisor Mateen Rizki, Ph.D. Chair, Department of Computer Science Committee on Final Examination John Gallagher, Ph.D. Thesis Advisor Mateen Rizki, Ph.D. Chair, Department of Computer Science Robert E. W. Fyffe, Ph.D. Vice President for Research and Dean of the Graduate School Abstract Coleman, Ashley.