The Theory Behind Overfitting, Cross Validation, Regularization, Bagging

Total Page:16

File Type:pdf, Size:1020Kb

The Theory Behind Overfitting, Cross Validation, Regularization, Bagging The Theory Behind Overfitting, Cross Validation, Regularization, Bagging, and Boosting: Tutorial Benyamin Ghojogh [email protected] Department of Electrical and Computer Engineering, Machine Learning Laboratory, University of Waterloo, Waterloo, ON, Canada Mark Crowley [email protected] Department of Electrical and Computer Engineering, Machine Learning Laboratory, University of Waterloo, Waterloo, ON, Canada Abstract 1. Introduction In this tutorial paper, we first define mean N Assume we have a dataset of instances f(xi; yi)gi=1 with squared error, variance, covariance, and bias d sample size N and dimensionality xi 2 R and yi 2 R. of both random variables and classifica- N N The fxigi=1 are the input data to the model and the fyigi=1 tion/predictor models. Then, we formulate the are the observations (labels). We denote the dataset by D true and generalization errors of the model so that N := jDj. This dataset is the union of the disjoint for both training and validation/test instances subsets, i.e., training set T and test set R; therefore: where we make use of the Stein’s Unbiased Risk Estimator (SURE). We define overfitting, under- D = T[R; (1) fitting, and generalization using the obtained true T\R = ?: (2) and generalization errors. We introduce cross validation and two well-known examples which For the training set, the observations (labels), yi’s, are are K-fold and leave-one-out cross validations. available. Although for the test set, we might also have We briefly introduce generalized cross validation yi’s, but we do not use them for training the model. The and then move on to regularization where we observations are continuous or come from a finite discrete use the SURE again. We work on both `2 and `1 set of values in classification and prediction (regression) norm regularizations. Then, we show that boot- tasks, respectively. Assume the sample size of training and strap aggregating (bagging) reduces the variance test sets are n := jT j and m := N − n, respectively; there- n of estimation. Boosting, specifically AdaBoost, fore, we have f(xi; yi)gi=1 as the training set. In some is introduced and it is explained as both an cases where we want to have validation set V as well, the additive model and a maximum margin model, datasets includes three disjoint subsets: i.e., Support Vector Machine (SVM). The upper bound on the generalization error of boosting D = T[R[V; (3) is also provided to show why boosting prevents T\R = ?; T\V = ?; V\R = ?: (4) from overfitting. As examples of regularization, arXiv:1905.12787v1 [stat.ML] 28 May 2019 the theory of ridge and lasso regressions, weight We will define the intuitions of training, test, and validation decay, noise injection to input/weights, and early sets later in this paper. stopping are explained. Random forest, dropout, In this paper, we introduce overfitting, cross validation, histogram of oriented gradients, and single shot generalized cross validation, regularization, bagging, and multi-box detector are explained as examples boosting and explain why they work theoretically. We also of bagging in machine learning and computer provide some examples of these methods in machine learn- vision. Finally, boosting tree and SVM models ing and computer vision. are mentioned as examples of boosting. 2. Mean Squared Error, Variance, and Bias 2.1. Measures for a Random Variable Assume we have variable X and we estimate it. Let the random variable Xb denote the estimate of X. The variance The Theory Behind Overfitting, Cross Validation, Regularization, Bagging, and Boosting: Tutorial 2 Figure 1. The dart example for (a) high bias and low variance, (b) low bias and high variance, (c) high bias and high variance, and (d) low bias and low variance. The worst and best cases are (c) and (d), respectively. The center of the circles is the true value of the variable. of estimating this random variable is defined as: The relation of MSE, variance, and bias is as follows: 2 MSE(Xb) = E (Xb − X) 2 Var(Xb) := E (Xb − E(Xb)) ; (5) 2 = E (Xb − E(Xb) + E(Xb) − X) 2 2 = E (Xb − E(Xb)) + (E(Xb) − X) which means average deviation of Xb from the mean of our + 2(Xb − E(Xb))(E(Xb) − X) estimate, E(Xb), where the deviation is squared for symme- (a) 2 2 try of difference. This variance can be restated as: = E (Xb − E(Xb)) + (E(Xb) − X) + 2 (E(Xb) − E(Xb))(E(Xb) − X) | {z } 2 2 0 Var(Xb) = E Xb + (E(Xb)) − 2XbE(Xb) (b) 2 (a) 2 2 = Var(Xb) + (Bias(Xb)) ; (9) = E(Xb ) + (E(Xb)) − 2E(Xb)E(Xb) 2 2 = E(Xb ) − (E(Xb)) ; (6) where (a) is because expectation is a linear operator and X and E(Xb) are not random, and (b) is because of Eqs. (5) and (7). where (a) is because expectation is a linear operator and If we have two random variables Xb and Yb, we can say: E(Xb) is not a random variable. Our estimation can have a bias. The bias of our estimate is (6) 2 2 Var(aXb + bYb) = E (aXb + bYb) − E(aXb + bYb) defined as: (a) 2 2 2 2 = a E(Xb ) + b E(Yb ) + 2ab E(XbYb) 2 2 2 2 − a (E(Xb)) − b (E(Yb)) − 2ab E(Yb)E(Yb) Bias(Xb) := E(Xb) − X; (7) (6) 2 2 = a Var(Xb) + b Var(Xb) + 2ab Cov(X;b Yb); (10) which means how much the mean of our estimate deviates where (a) is because of linearity of expectation and the from the original X. Cov(X;b Yb) is covariance defined as: The Mean Squared Error (MSE) of our estimate, Xb, is de- fined as: Cov(X;b Yb) := E(XbYb) − E(Yb) E(Yb): (11) If the two random variables are independent, i.e., X ?? Y , 2 MSE(Xb) := E (Xb − X) ; (8) we have: (a) ZZ ?? ZZ E(XbYb) = xyf(x; y)dxdy = xyf(x)f(y)dxdy which means how much our estimate deviates from the bb b b b b bb b b b b original X. Z Z Z = yf(y) xf(x)dx dy = E(Xb) yf(y)dy The intuition of bias, variance, and MSE is illustrated in b b b b b b b b b Fig.1 where the estimations are like a dart game. We have | {z } | {z } E(Xb) E(Yb ) four cases with low/high values of bias and variance which are depicted in this figure. = E(Xb) E(Yb) =) Cov(X;b Yb) = 0; (12) The Theory Behind Overfitting, Cross Validation, Regularization, Bagging, and Boosting: Tutorial 3 where (a) is according to definition of expectation. Note that Eq. (12) is not true for the reverse implication (we can prove by counterexample). We can extend Eqs. (10) and (11) to multiple random vari- ables: k X Var aiXi i=1 k k k X 2 X X = ai Var(Xi) + aiajCov(Xi;Xj); i=1 i=1 j=1;j6=i (13) k k k k X1 X2 X1 X2 Cov aiXi; bjYj = ai bj Cov(Xi;Yj); Figure 2. The true model and the estimated model which is trained i=1 j=1 i=1 j=1 using the input training data and their observations. The observa- (14) tions are obtained from the outputs of the true model fed with the training input but corrupted by the noise. After the model is where ai’s and bj’s are not random. trained, it can be used to estimate the observation for either train- ing or test input data. 2.2. Measures for a Model Assume we have a function f which gets the i-th input xi and outputs fi = f(xi). Figure2 shows this function and its input and output. We wish to know the function which we call it the true model but we do not have access to it as it is unknown. Also, the pure outputs (true observations), fi’s, are not available. The output may be corrupted with an additive noise "i: yi = fi + "i; (15) 2 where the noise is "i ∼ N (0; σ ). Therefore: 2 (6) 2 2 Figure 3. The triangle of variance, bias, and MSE. The f and fb ("i) = 0; (" ) = ar("i) + ( ("i)) = σ ; (16) E E i V E are the true and estimated models, respectively. The true observation fi is not random, thus: the estimation is a member of a discrete set of possible ob- E(fi) = fi: (17) servations. n The input training data fxigi=1 and their corrupted obser- The definitions of variance, bias, and MSE, i.e., Eqs. (5), vations fy gn are available to us. We would like to ap- i i=1 (7), and (8), can also be used for the estimation fbi of the proximate (estimate) the true model by a model fb in or- true model fi. The Eq. (9) can be illustrated for the model n der to estimate the observations fyigi=1 from the input f as in Fig.3 which holds because of Pythagorean theorem. n n fxigi=1. Calling the estimated observations by fybigi=1, n n 2.3. Measures for Ensemble of Models we want the fybigi=1 to be as close as possible to fyigi=1 n for the training input data fxigi=1. We train the model us- If we have an ensemble of models (Polikar, 2012), we can ing the training data in order to estimate the true model. have some similar definitions of bias and variance (e.g., see After training the model, it can be used to estimate the out- the Appendix C in (Schapire et al., 1998)).
Recommended publications
  • Harmless Overfitting: Using Denoising Autoencoders in Estimation of Distribution Algorithms
    Journal of Machine Learning Research 21 (2020) 1-31 Submitted 10/16; Revised 4/20; Published 4/20 Harmless Overfitting: Using Denoising Autoencoders in Estimation of Distribution Algorithms Malte Probst∗ [email protected] Honda Research Institute EU Carl-Legien-Str. 30, 63073 Offenbach, Germany Franz Rothlauf [email protected] Information Systems and Business Administration Johannes Gutenberg-Universit¨atMainz Jakob-Welder-Weg 9, 55128 Mainz, Germany Editor: Francis Bach Abstract Estimation of Distribution Algorithms (EDAs) are metaheuristics where learning a model and sampling new solutions replaces the variation operators recombination and mutation used in standard Genetic Algorithms. The choice of these models as well as the correspond- ing training processes are subject to the bias/variance tradeoff, also known as under- and overfitting: simple models cannot capture complex interactions between problem variables, whereas complex models are susceptible to modeling random noise. This paper suggests using Denoising Autoencoders (DAEs) as generative models within EDAs (DAE-EDA). The resulting DAE-EDA is able to model complex probability distributions. Furthermore, overfitting is less harmful, since DAEs overfit by learning the identity function. This over- fitting behavior introduces unbiased random noise into the samples, which is no major problem for the EDA but just leads to higher population diversity. As a result, DAE-EDA runs for more generations before convergence and searches promising parts of the solution space more thoroughly. We study the performance of DAE-EDA on several combinatorial single-objective optimization problems. In comparison to the Bayesian Optimization Al- gorithm, DAE-EDA requires a similar number of evaluations of the objective function but is much faster and can be parallelized efficiently, making it the preferred choice especially for large and difficult optimization problems.
    [Show full text]
  • More Perceptron
    More Perceptron Instructor: Wei Xu Some slides adapted from Dan Jurfasky, Brendan O’Connor and Marine Carpuat A Biological Neuron https://www.youtube.com/watch?v=6qS83wD29PY Perceptron (an artificial neuron) inputs Perceptron Algorithm • Very similar to logistic regression • Not exactly computing gradient (simpler) vs. weighted sum of features Online Learning • Rather than making a full pass through the data, compute gradient and update parameters after each training example. Stochastic Gradient Descent • Updates (or more specially, gradients) will be less accurate, but the overall effect is to move in the right direction • Often works well and converges faster than batch learning Online Learning • update parameters for each training example Initialize weight vector w = 0 Create features Loop for K iterations Loop for all training examples x_i, y_i … update_weights(w, x_i, y_i) Perceptron Algorithm • Very similar to logistic regression • Not exactly computing gradient Initialize weight vector w = 0 Loop for K iterations Loop For all training examples x_i if sign(w * x_i) != y_i Error-driven! w += y_i * x_i Error-driven Intuition • For a given example, makes a prediction, then checks to see if this prediction is correct. - If the prediction is correct, do nothing. - If the prediction is wrong, change its parameters so that it would do better on this example next time around. Error-driven Intuition Error-driven Intuition Error-driven Intuition Exercise Perceptron (vs. LR) • Only hyperparameter is maximum number of iterations (LR also needs learning rate) • Guaranteed to converge if the data is linearly separable (LR always converge) objective function is convex What if non linearly separable? • In real-world problems, this is nearly always the case.
    [Show full text]
  • Measuring Overfitting and Mispecification in Nonlinear Models
    WP 11/25 Measuring overfitting and mispecification in nonlinear models Marcel Bilger Willard G. Manning August 2011 york.ac.uk/res/herc/hedgwp Measuring overfitting and mispecification in nonlinear models Marcel Bilger, Willard G. Manning The Harris School of Public Policy Studies, University of Chicago, USA Abstract We start by proposing a new measure of overfitting expressed on the untransformed scale of the dependent variable, which is generally the scale of interest to the analyst. We then show that with nonlinear models shrinkage due to overfitting gets confounded by shrinkage—or expansion— arising from model misspecification. Out-of-sample predictive calibration can in fact be expressed as in-sample calibration times 1 minus this new measure of overfitting. We finally argue that re-calibration should be performed on the scale of interest and provide both a simulation study and a real-data illustration based on health care expenditure data. JEL classification: C21, C52, C53, I11 Keywords: overfitting, shrinkage, misspecification, forecasting, health care expenditure 1. Introduction When fitting a model, it is well-known that we pick up part of the idiosyncratic char- acteristics of the data as well as the systematic relationship between a dependent and explanatory variables. This phenomenon is known as overfitting and generally occurs when a model is excessively complex relative to the amount of data available. Overfit- ting is a major threat to regression analysis in terms of both inference and prediction. When models greatly over-explain the data at hand they can even show relations which reflect chance only. Consequently, overfitting casts doubt on the true statistical signif- icance of the effects found by the analyst as well as the magnitude of the response.
    [Show full text]
  • Early Stopping for Kernel Boosting Algorithms: a General Analysis with Localized Complexities
    IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 65, NO. 10, OCTOBER 2019 6685 Early Stopping for Kernel Boosting Algorithms: A General Analysis With Localized Complexities Yuting Wei , Fanny Yang, and Martin J. Wainwright, Senior Member, IEEE Abstract— Early stopping of iterative algorithms is a widely While the basic idea of early stopping is fairly old (e.g., [1], used form of regularization in statistics, commonly used in [33], [37]), recent years have witnessed renewed interests in its conjunction with boosting and related gradient-type algorithms. properties, especially in the context of boosting algorithms and Although consistency results have been established in some settings, such estimators are less well-understood than their neural network training (e.g., [13], [27]). Over the past decade, analogues based on penalized regularization. In this paper, for a a line of work has yielded some theoretical insight into early relatively broad class of loss functions and boosting algorithms stopping, including works on classification error for boosting (including L2-boost, LogitBoost, and AdaBoost, among others), algorithms [4], [14], [19], [25], [41], [42], L2-boosting algo- we exhibit a direct connection between the performance of a rithms for regression [8], [9], and similar gradient algorithms stopped iterate and the localized Gaussian complexity of the associated function class. This connection allows us to show that in reproducing kernel Hilbert spaces (e.g. [11], [12], [28], the local fixed point analysis of Gaussian or Rademacher com- [36], [41]). A number of these papers establish consistency plexities, now standard in the analysis of penalized estimators, results for particular forms of early stopping, guaranteeing can be used to derive optimal stopping rules.
    [Show full text]
  • Over-Fitting in Model Selection with Gaussian Process Regression
    Over-Fitting in Model Selection with Gaussian Process Regression Rekar O. Mohammed and Gavin C. Cawley University of East Anglia, Norwich, UK [email protected], [email protected] http://theoval.cmp.uea.ac.uk/ Abstract. Model selection in Gaussian Process Regression (GPR) seeks to determine the optimal values of the hyper-parameters governing the covariance function, which allows flexible customization of the GP to the problem at hand. An oft-overlooked issue that is often encountered in the model process is over-fitting the model selection criterion, typi- cally the marginal likelihood. The over-fitting in machine learning refers to the fitting of random noise present in the model selection criterion in addition to features improving the generalisation performance of the statistical model. In this paper, we construct several Gaussian process regression models for a range of high-dimensional datasets from the UCI machine learning repository. Afterwards, we compare both MSE on the test dataset and the negative log marginal likelihood (nlZ), used as the model selection criteria, to find whether the problem of overfitting in model selection also affects GPR. We found that the squared exponential covariance function with Automatic Relevance Determination (SEard) is better than other kernels including squared exponential covariance func- tion with isotropic distance measure (SEiso) according to the nLZ, but it is clearly not the best according to MSE on the test data, and this is an indication of over-fitting problem in model selection. Keywords: Gaussian process · Regression · Covariance function · Model selection · Over-fitting 1 Introduction Supervised learning tasks can be divided into two main types, namely classifi- cation and regression problems.
    [Show full text]
  • On Overfitting and Asymptotic Bias in Batch Reinforcement Learning With
    Journal of Artificial Intelligence Research 65 (2019) 1-30 Submitted 06/2018; published 05/2019 On Overfitting and Asymptotic Bias in Batch Reinforcement Learning with Partial Observability Vincent François-Lavet [email protected] Guillaume Rabusseau [email protected] Joelle Pineau [email protected] School of Computer Science, McGill University University Street 3480, Montreal, QC, H3A 2A7, Canada Damien Ernst [email protected] Raphael Fonteneau [email protected] Montefiore Institute, University of Liege Allée de la découverte 10, 4000 Liège, Belgium Abstract This paper provides an analysis of the tradeoff between asymptotic bias (suboptimality with unlimited data) and overfitting (additional suboptimality due to limited data) in the context of reinforcement learning with partial observability. Our theoretical analysis formally characterizes that while potentially increasing the asymptotic bias, a smaller state representation decreases the risk of overfitting. This analysis relies on expressing the quality of a state representation by bounding L1 error terms of the associated belief states. Theoretical results are empirically illustrated when the state representation is a truncated history of observations, both on synthetic POMDPs and on a large-scale POMDP in the context of smartgrids, with real-world data. Finally, similarly to known results in the fully observable setting, we also briefly discuss and empirically illustrate how using function approximators and adapting the discount factor may enhance the tradeoff between asymptotic bias and overfitting in the partially observable context. 1. Introduction This paper studies sequential decision-making problems that may be modeled as Markov Decision Processes (MDP) but for which the state is partially observable.
    [Show full text]
  • Overfitting Princeton University COS 495 Instructor: Yingyu Liang Review: Machine Learning Basics Math Formulation
    Machine Learning Basics Lecture 6: Overfitting Princeton University COS 495 Instructor: Yingyu Liang Review: machine learning basics Math formulation • Given training data 푥푖, 푦푖 : 1 ≤ 푖 ≤ 푛 i.i.d. from distribution 퐷 1 • Find 푦 = 푓(푥) ∈ 퓗 that minimizes 퐿෠ 푓 = σ푛 푙(푓, 푥 , 푦 ) 푛 푖=1 푖 푖 • s.t. the expected loss is small 퐿 푓 = 피 푥,푦 ~퐷[푙(푓, 푥, 푦)] Machine learning 1-2-3 • Collect data and extract features • Build model: choose hypothesis class 퓗 and loss function 푙 • Optimization: minimize the empirical loss Machine learning 1-2-3 Feature mapping Maximum Likelihood • Collect data and extract features • Build model: choose hypothesis class 퓗 and loss function 푙 • Optimization: minimize the empirical loss Occam’s razor Gradient descent; convex optimization Overfitting Linear vs nonlinear models 2 푥1 2 푥2 푥1 2푥1푥2 푦 = sign(푤푇휙(푥) + 푏) 2푐푥1 푥2 2푐푥2 푐 Polynomial kernel Linear vs nonlinear models • Linear model: 푓 푥 = 푎0 + 푎1푥 2 3 푀 • Nonlinear model: 푓 푥 = 푎0 + 푎1푥 + 푎2푥 + 푎3푥 + … + 푎푀 푥 • Linear model ⊆ Nonlinear model (since can always set 푎푖 = 0 (푖 > 1)) • Looks like nonlinear model can always achieve same/smaller error • Why one use Occam’s razor (choose a smaller hypothesis class)? Example: regression using polynomial curve 푡 = sin 2휋푥 + 휖 Figure from Machine Learning and Pattern Recognition, Bishop Example: regression using polynomial curve 푡 = sin 2휋푥 + 휖 Regression using polynomial of degree M Figure from Machine Learning and Pattern Recognition, Bishop Example: regression using polynomial curve 푡 = sin 2휋푥 + 휖 Figure from Machine Learning
    [Show full text]
  • Search for Ultra-High Energy Photons with the Pierre Auger Observatory Using Deep Learning Techniques
    Search for Ultra-High Energy Photons with the Pierre Auger Observatory using Deep Learning Techniques von Tobias Alexander Pan Masterarbeit in Physik vorgelegt der Fakult¨atf¨urMathematik, Informatik und Naturwissenschaften der RWTH Aachen im Februar 2020 angefertigt im III. Physikalischen Institut A bei Jun.-Prof. Dr. Thomas Bretz Prof. Dr. Thomas Hebbeker Contents 1 Introduction1 2 Cosmic rays3 2.1 Cosmic ray physics................................3 2.2 Cosmic ray-induced extensive air showers...................6 2.3 Attributes of photon-induced air showers....................8 2.4 Detection principles............................... 11 3 The Pierre Auger Observatory 13 3.1 The Surface Detector............................... 13 3.2 The Fluorescence Detector............................ 15 3.3 The Infill array.................................. 16 3.4 AugerPrime Upgrade............................... 17 3.5 Standard event reconstruction.......................... 18 4 Simulation 21 4.1 CORSIKA..................................... 21 4.2 Offline Software Framework........................... 22 4.3 Data set selection................................. 23 4.4 Preprocessing................................... 25 5 Photon search with machine learning 31 6 Deep neural networks 37 6.1 Basic principles.................................. 37 6.2 Regularization.................................. 40 6.3 Convolutional networks............................. 41 7 Photon search using deep neural networks 43 7.1 Network architecture..............................
    [Show full text]
  • A Robust Hybrid of Lasso and Ridge Regression
    A robust hybrid of lasso and ridge regression Art B. Owen Stanford University October 2006 Abstract Ridge regression and the lasso are regularized versions of least squares regression using L2 and L1 penalties respectively, on the coefficient vector. To make these regressions more robust we may replace least squares with Huber’s criterion which is a hybrid of squared error (for relatively small errors) and absolute error (for relatively large ones). A reversed version of Huber’s criterion can be used as a hybrid penalty function. Relatively small coefficients contribute their L1 norm to this penalty while larger ones cause it to grow quadratically. This hybrid sets some coefficients to 0 (as lasso does) while shrinking the larger coefficients the way ridge regression does. Both the Huber and reversed Huber penalty functions employ a scale parameter. We provide an objective function that is jointly convex in the regression coefficient vector and these two scale parameters. 1 Introduction We consider here the regression problem of predicting y ∈ R based on z ∈ Rd. The training data are pairs (zi, yi) for i = 1, . , n. We suppose that each vector p of predictor vectors zi gets turned into a feature vector xi ∈ R via zi = φ(xi) for some fixed function φ. The predictor for y is linear in the features, taking the form µ + x0β where β ∈ Rp. In ridge regression (Hoerl and Kennard, 1970) we minimize over β, a criterion of the form n p X 0 2 X 2 (yi − µ − xiβ) + λ βj , (1) i=1 j=1 for a ridge parameter λ ∈ [0, ∞].
    [Show full text]
  • Using Validation Sets to Avoid Overfitting in Adaboost
    Using Validation Sets to Avoid Overfitting in AdaBoost ∗ Tom Bylander and Lisa Tate Department of Computer Science, University of Texas at San Antonio, San Antonio, TX 78249 USA {bylander, ltate}@cs.utsa.edu Abstract In both cases, the training set is used to find a small hy- pothesis space, and the validation set is used to select a hy- AdaBoost is a well known, effective technique for increas- pothesis from that space. The justification is that in a large ing the accuracy of learning algorithms. However, it has the potential to overfit the training set because its objective is hypothesis space, minimizing training error will often result to minimize error on the training set. We demonstrate that in overfitting. The training set can instead be used to estab- overfitting in AdaBoost can be alleviated in a time-efficient lish a set of choices as training error is minimized, with the manner using a combination of dagging and validation sets. validation set used to reverse or modify those choices. Half of the training set is removed to form the validation set. A validation set could simply be used for early stopping The sequence of base classifiers, produced by AdaBoost from in AdaBoost. However, we obtain the most success by also the training set, is applied to the validation set, creating a performing 2-dagging (Ting & Witten 1997), and modifying modified set of weights. The training and validation sets are weights. Dagging is disjoint aggregation, where a training switched, and a second pass is performed. The final classi- set is partitioned into subsets and training is performed on fier votes using both sets of weights.
    [Show full text]
  • Lasso Reference Manual Release 17
    STATA LASSO REFERENCE MANUAL RELEASE 17 ® A Stata Press Publication StataCorp LLC College Station, Texas Copyright c 1985–2021 StataCorp LLC All rights reserved Version 17 Published by Stata Press, 4905 Lakeway Drive, College Station, Texas 77845 Typeset in TEX ISBN-10: 1-59718-337-7 ISBN-13: 978-1-59718-337-6 This manual is protected by copyright. All rights are reserved. No part of this manual may be reproduced, stored in a retrieval system, or transcribed, in any form or by any means—electronic, mechanical, photocopy, recording, or otherwise—without the prior written permission of StataCorp LLC unless permitted subject to the terms and conditions of a license granted to you by StataCorp LLC to use the software and documentation. No license, express or implied, by estoppel or otherwise, to any intellectual property rights is granted by this document. StataCorp provides this manual “as is” without warranty of any kind, either expressed or implied, including, but not limited to, the implied warranties of merchantability and fitness for a particular purpose. StataCorp may make improvements and/or changes in the product(s) and the program(s) described in this manual at any time and without notice. The software described in this manual is furnished under a license agreement or nondisclosure agreement. The software may be copied only in accordance with the terms of the agreement. It is against the law to copy the software onto DVD, CD, disk, diskette, tape, or any other medium for any purpose other than backup or archival purposes. The automobile dataset appearing on the accompanying media is Copyright c 1979 by Consumers Union of U.S., Inc., Yonkers, NY 10703-1057 and is reproduced by permission from CONSUMER REPORTS, April 1979.
    [Show full text]
  • Overfitting Can Be Harmless for Basis Pursuit, but Only to a Degree
    Overfitting Can Be Harmless for Basis Pursuit, But Only to a Degree Peizhong Ju Xiaojun Lin Jia Liu School of ECE School of ECE Department of ECE Purdue University Purdue University The Ohio State University West Lafayette, IN 47906 West Lafayette, IN 47906 Columbus, OH 43210 [email protected] [email protected] [email protected] Abstract Recently, there have been significant interests in studying the so-called “double- descent” of the generalization error of linear regression models under the overpa- rameterized and overfitting regime, with the hope that such analysis may provide the first step towards understanding why overparameterized deep neural networks (DNN) still generalize well. However, to date most of these studies focused on the min `2-norm solution that overfits the data. In contrast, in this paper we study the overfitting solution that minimizes the `1-norm, which is known as Basis Pursuit (BP) in the compressed sensing literature. Under a sparse true linear regression model with p i.i.d. Gaussian features, we show that for a large range of p up to a limit that grows exponentially with the number of samples n, with high probability the model error of BP is upper bounded by a value that decreases with p. To the best of our knowledge, this is the first analytical result in the literature establishing the double-descent of overfitting BP for finite n and p. Further, our results reveal significant differences between the double-descent of BP and min `2-norm solu- tions. Specifically, the double-descent upper-bound of BP is independent of the signal strength, and for high SNR and sparse models the descent-floor of BP can be much lower and wider than that of min `2-norm solutions.
    [Show full text]