Introduction to Empirical Processes and Semiparametric Inference Lecture 04: Overview Continued

Total Page:16

File Type:pdf, Size:1020Kb

Introduction to Empirical Processes and Semiparametric Inference Lecture 04: Overview Continued Empirical Processes: Lecture 04 Spring, 2014 Introduction to Empirical Processes and Semiparametric Inference Lecture 04: Overview Continued Michael R. Kosorok, Ph.D. Professor and Chair of Biostatistics Professor of Statistics and Operations Research University of North Carolina-Chapel Hill 1 Empirical Processes: Lecture 04 Spring, 2014 We say that a map φ : Dφ ⊂ D 7! E, is Gateaux-differentiableˆ at θ 2 Dφ 0 if there exists a map φθ : D 7! E such that φ(θ + th − φ(θ) 0 − φ (h) ! 0 ; t θ where θ + th 2 Dφ for every t > 0 small enough. 2 Empirical Processes: Lecture 04 Spring, 2014 We say that the map is Hadamard-differentiable if there exists a 0 continuous linear map φθ : D 7! E such that φ(θ + th − φ(θ) 0 sup φθ(h) ! 0 ; (1) h2K t for every compact set K ⊂ D where θ + th 2 Dφ for every t > 0 small enough. We say that the map is Frechet-differentiable´ if the condition (1) holds for every bounded subset K ⊂ D. 3 Empirical Processes: Lecture 04 Spring, 2014 §Semiparametrics: Models and Efficiency ¤ ¦ ¥ A statistical model is a collection of probability measures fP 2 Pg that specify the distribution of a random observation X. Consider the linear model Y = β0Z + e, where the observed data is k X = (Y; Z), β 2 R is unknown, and the joint distribution of (Z; e) satisfies • E(ejZ) = 0 almost surely, • E(e2jZ) ≤ K < 1 almost surely, and • the distribution of (Z; e) is otherwise unspecified (mostly). 4 Empirical Processes: Lecture 04 Spring, 2014 In this instance, the model P has • an unkown parameter β of interest and • some other unknown components of secondary interest but which allow flexibility (the joint distribution of (Z; e)). In general, a model P has • A parameter (often denoted ) of interest and • several partially unspecified components not of interest. 5 Empirical Processes: Lecture 04 Spring, 2014 For semiparametric models, = (θ; η), where • θ is finite dimensional and usually of primary interest, and • η is infinite dimensional and usually of secondary interest. • There may be other components not of interest which are not parameterized at all. • θ and η may have multiple subcomponents. • Choices of parameter names are contextual and vary in the literature, although there are some consistencies. 6 Empirical Processes: Lecture 04 Spring, 2014 The goals of semiparametric inference are primarily to: • Select an appropriate model for inference on X; • Estimate one or more subcomponents of , sometimes θ alone is the focus; • Conduct inference (e.g., confidence intervals) for the parameters of interest; • Try to obtain efficient estimators. The standard set-up for inference we will use in this book is where estimation and inference is based on a sample, X1;:::;Xn of i.i.d. realizations of a distribution P 2 P. 7 Empirical Processes: Lecture 04 Spring, 2014 The generic prameter of interest can be viewed as a function of the distribution (P ), since if we know the distribution, we know the parameter value. An estimator Tn (based on the data X1;:::;Xn) is efficient for (P ) if p the limiting variance V of n(Tn − (P )) is the smallest among all regular estimators of (P ). The inverse of V is the information for Tn. 8 Empirical Processes: Lecture 04 Spring, 2014 The optimal efficiency of a parameter (P ) depends on the complexity of the model P. Estimation under P is more taxing than estimation under any parametric submodel P0 = fPθ : θ 2 Θ0g ⊂ P, where Θ0 is finite-dimensional. For example, in the linear regression model P setting above, if we assume Z and e are independent and e is N(0; σ2), where σ2 is unknown, and call this model P0, P0 is a parametric submodel of P. 9 Empirical Processes: Lecture 04 Spring, 2014 Thus, and in general, information for P is worse (less) than information for any parametric submodel P0 ⊂ P. For a semiparametric model P, if the information for the regular estimator Tn (to estimate (P )) equals the minimum (infemum) of the Fisher informations for all parametric submodels of P, then Tn is semiparametric efficient for (P ), and the associated information for Tn is called the “efficient information” (also called the efficient information for P). This is because the only settings with more information than Tn are for parametric (not semiparametric) submodels. 10 Empirical Processes: Lecture 04 Spring, 2014 Interestingly, when (P ) is one-dimensional, it is always possible to identify a one-dimensional parametric submodel that has the same Fisher information as the efficient information for P. A parametric submodel that achieves the efficient information is called a least favorable or hardest submodel. Fortunately, finding the efficient information for (P ) only requires consideration of one-dimensional parametric submodels fPt : t 2 [0; )g surrounding representative (true) distributions P 2 P, where P0 = P and Pt 2 P for all t 2 [0; ). 11 Empirical Processes: Lecture 04 Spring, 2014 If P has a dominating measure µ, then each P can be expressed as a density p. In this case, we require the densities around p to be smooth enough so R 2 that g(x) = @pt(x)=(@t)jt=0 exists with X g (x)p(x)µ(dx) < 1. More generally, we say that a submodel is differentiable in quadratic mean with score function g : X 7! R if 2 Z " 1=2 1=2 # (dPt(x)) − (dP (x)) 1 − g(x)(dP (x))1=2 ! 0: (2) t 2 12 Empirical Processes: Lecture 04 Spring, 2014 Sometimes it is important for us to consider a collection of many one-dimensional submodels surrounding a representative P , each submodel represented by a score function g: such a collection of score functions P_P is called a tangent set. 2 Because, for any g 2 P_P , P g = 0 and P g < 1, these tangent sets 0 are subsets of L2(P ), the space of all function h : X 7! R with P h = 0 and P h2 < 1. When the tangent set is closed under linear combinations, it is called a tangent space. 13 Empirical Processes: Lecture 04 Spring, 2014 Consider a parametric model P = fPθ : θ 2 Θg, where Θ is an open k subset of R and P is dominated by µ, such that the classical score _ _ 2 function `θ(x) = @ log pθ(x)=(@θ) exists with Pθk`θk < 1 (and a few other conditions hold). One can show that each of the one-dimensional submodels 0 _ k g(x) = h `θ(x), for any h 2 R , satisfies (2), forming the tangent space _ 0 _ k PPθ = fh `θ : h 2 R g. Thus there is a simple and direct connection between the classical score function and the more general tangent spaces. 14 Empirical Processes: Lecture 04 Spring, 2014 Continuing with the parametric setting, we say that an estimator θ^ of θ is efficient for estimating θ if • it is regular and _ _0 • it’s information achieves the Cramer-Rao` lower bound P [`θ`θ]. Thus the tangent set for the model contains information about the optimal efficiency. This is also true for semiparametric models, although the relationship between tangents sets and the optimal information is more complex. 15 Empirical Processes: Lecture 04 Spring, 2014 k Consider estimation of the parameter (P ) 2 R for the model P: For any estimator Tn of (P ), if p p ˇ n(Tn − (P )) = nPn P + oP (1); then ˇ • P is an influence function for (P ) and • Tn is asymptotically linear. 16 Empirical Processes: Lecture 04 Spring, 2014 For a given tangent set P_P , assume for each submodel fPt : 0 ≤ t < g satisfying (2) with some g 2 P_P and some > 0, that @ (P ) t _ = P (g); @t t=0 _ 0 k for some linear map P : L2(P ) 7! R : In this setting, we say that is differentiable at P relative to P_P . 17 Empirical Processes: Lecture 04 Spring, 2014 When P_P is a linear space, there exists a measurable function ~ k P : X 7! R such that _ h ~ i P (g) = P P (X)g(X) ; for each g 2 P_P . ~ _ 0 The function P 2 PP ⊂ L2(P ) is unique and is called the efficient influence funtion for the parameter in the model P. 18 Empirical Processes: Lecture 04 Spring, 2014 ~ The efficient influence function P for can usually be found by taking ˇ any influence function P and projecting it onto P_P . Moreover, we will see in Theorem 18.8 that full efficiency of an estimator Tn of (P ) can be verified by checking that the influence function of Tn lies in P_P . 19 Empirical Processes: Lecture 04 Spring, 2014 k Consider a parametric model P = fPθ : θ 2 Θg, where Θ ⊂ R . k d Suppose the parameter of interest is (Pθ) = f(θ), for f : R 7! R _ with derivative fθ at θ, and that the Fisher information for θ, h i _ _0 Iθ ≡ P `θ`θ is invertible. 0 _ k Recall that, in this case, g(X) = h `θ(X), for some h 2 R , and thus _ _ Pt = Pθ+th, which implies that P (g) = fθh. 20 Empirical Processes: Lecture 04 Spring, 2014 Since also we have h i h i _ −1 _ _ −1 _ _0 fθIθ P `θ(X)g(X) = fθIθ P `θ`θh _ = fθh; h i _ _ ~ we have that P (g) = fθh = P P g , provided ~ _ −1 _ P (X) = fθIθ `θ(X): p p ~ Any estimator Tn for which n(Tn − (P )) = nPn P + oP (1) has _ −1 _T asymptotic variance fθIθ fθ and thus achieves the Cramer` Rao lower bound and is efficient.
Recommended publications
  • Instrumental Regression in Partially Linear Models
    10-167 Research Group: Econometrics and Statistics September, 2009 Instrumental Regression in Partially Linear Models JEAN-PIERRE FLORENS, JAN JOHANNES AND SEBASTIEN VAN BELLEGEM INSTRUMENTAL REGRESSION IN PARTIALLY LINEAR MODELS∗ Jean-Pierre Florens1 Jan Johannes2 Sebastien´ Van Bellegem3 First version: January 24, 2008. This version: September 10, 2009 Abstract We consider the semiparametric regression Xtβ+φ(Z) where β and φ( ) are unknown · slope coefficient vector and function, and where the variables (X,Z) are endogeneous. We propose necessary and sufficient conditions for the identification of the parameters in the presence of instrumental variables. We also focus on the estimation of β. An incorrect parameterization of φ may generally lead to an inconsistent estimator of β, whereas even consistent nonparametric estimators for φ imply a slow rate of convergence of the estimator of β. An additional complication is that the solution of the equation necessitates the inversion of a compact operator that has to be estimated nonparametri- cally. In general this inversion is not stable, thus the estimation of β is ill-posed. In this paper, a √n-consistent estimator for β is derived under mild assumptions. One of these assumptions is given by the so-called source condition that is explicitly interprated in the paper. Finally we show that the estimator achieves the semiparametric efficiency bound, even if the model is heteroscedastic. Monte Carlo simulations demonstrate the reasonable performance of the estimation procedure on finite samples. Keywords: Partially linear model, semiparametric regression, instrumental variables, endo- geneity, ill-posed inverse problem, Tikhonov regularization, root-N consistent estimation, semiparametric efficiency bound JEL classifications: Primary C14; secondary C30 ∗We are grateful to R.
    [Show full text]
  • Arxiv:1609.06421V4 [Math.ST] 26 Sep 2019 Semiparametric Identification and Fisher Information∗
    Semiparametric Identification and Fisher Information∗ Juan Carlos Escanciano† Universidad Carlos III de Madrid September 25th, 2019 Abstract This paper provides a systematic approach to semiparametric identification that is based on statistical information as a measure of its “quality”. Identification can be regular or irregular, depending on whether the Fisher information for the parameter is positive or zero, respectively. I first characterize these cases in models with densities linear in a nonparametric parameter. I then introduce a novel “generalized Fisher information”. If positive, it implies (possibly irregular) identification when other conditions hold. If zero, it implies impossibility results on rates of estimation. Three examples illustrate the applicability of the general results. First, I find necessary conditions for semiparametric regular identification in a structural model for unemployment duration with two spells and nonparametric heterogeneity. Second, I show irregular identification of the median willingness to pay in contingent valuation studies. Finally, I study identification of the discount factor and average measures of risk aversion in a nonparametric Euler Equation with nonparametric measurement error in consumption. Keywords: Identification; Semiparametric Models; Fisher Information. arXiv:1609.06421v4 [math.ST] 26 Sep 2019 JEL classification: C14; C31; C33; C35 ∗First version: September 20th, 2016. †Department of Economics, Universidad Carlos III de Madrid, email: [email protected]. Research funded by the Spanish Programa de Generaci´on de Conocimiento, reference number PGC2018-096732-B-I00. I thank Michael Jansson, Ulrich M¨uller, Whitney Newey, Jack Porter, Pedro Sant’Anna, Ruli Xiao, and seminar par- ticipants at BC, Indiana, MIT, Texas A&M, UBC, Vanderbilt and participants of the 2018 Conference on Identification in Econometrics for useful comments.
    [Show full text]
  • Semiparametric Efficiency
    Faculteit Wetenschappen Vakgroep Toegepaste Wiskunde en Informatica Semiparametric Efficiency Karel Vermeulen Prof. Dr. S. Vansteelandt Proefschrift ingediend tot het behalen van de graad van Master in de Wiskunde, afstudeerrichting Toegepaste Wiskunde Academiejaar 2010-2011 To my parents, my brother Lukas To my best friend Sara A mathematical theory is not to be considered complete until you have made it so clear that you can explain it to the first man whom you meet on the street::: David Hilbert Preface My interest in semiparametric theory awoke several years ago, two and a half to be precise. That time, I was in my third year of mathematics. I had to choose a subject for my Bachelor thesis. My decision was: A geometrical approach to the asymptotic efficiency of estimators, based on the monograph by Anastasios Tsiatis, [35], under the supervision of Prof. Dr. Stijn Vansteelandt. In this manner, I entered the world of semiparametric efficiency. However, at that point I did not know it was just the beginning. Shortly after I wrote this Bachelor thesis, Prof. Dr. Stijn Vansteelandt asked me to be involved in research on semiparametric inference for so called probabilistic index models, in the context of a one month student job. Under his guidance, I applied semiparametric estimation theory to this setting and obtained the semiparametric efficiency bound to which efficiency of estimators in this model can be measured. Some results of this research are contained within this thesis. While short, this experience really convinced me I wanted to write a thesis in semiparametric efficiency. That feeling was only more encouraged after following the course Causality and Missing Data, taught by Prof.
    [Show full text]
  • Efficient Estimation in Single Index Models Through Smoothing Splines
    arXiv: 1612.00068 Efficient Estimation in Single Index Models through Smoothing splines Arun K. Kuchibhotla and Rohit K. Patra University of Pennsylvania and University of Florida e-mail: [email protected]; [email protected] Abstract: We consider estimation and inference in a single index regression model with an unknown but smooth link function. In contrast to the standard approach of using kernels or regression splines, we use smoothing splines to estimate the smooth link function. We develop a method to compute the penalized least squares estimators (PLSEs) of the para- metric and the nonparametric components given independent and identically distributed (i.i.d.) data. We prove the consistency and find the rates of convergence of the estimators. We establish asymptotic normality under under mild assumption and prove asymptotic effi- ciency of the parametric component under homoscedastic errors. A finite sample simulation corroborates our asymptotic theory. We also analyze a car mileage data set and a Ozone concentration data set. The identifiability and existence of the PLSEs are also investigated. Keywords and phrases: least favorable submodel, penalized least squares, semiparamet- ric model. 1. Introduction Consider a regression model where one observes i.i.d. copies of the predictor X ∈ Rd and the response Y ∈ R and is interested in estimating the regression function E(Y |X = ·). In nonpara- metric regression E(Y |X = ·) is generally assumed to satisfy some smoothness assumptions (e.g., twice continuously differentiable), but no assumptions are made on the form of dependence on X. While nonparametric models offer flexibility in modeling, the price for this flexibility can be high for two main reasons: the estimation precision decreases rapidly as d increases (“curse of dimensionality”) and the estimator can be hard to interpret when d > 1.
    [Show full text]
  • Introduction to Empirical Processes and Semiparametric Inference1
    This is page i Printer: Opaque this Introduction to Empirical Processes and Semiparametric Inference1 Michael R. Kosorok August 2006 1c 2006 SPRINGER SCIENCE+BUSINESS MEDIA, INC. All rights re- served. Permission is granted to print a copy of this preliminary version for non- commercial purposes but not to distribute it in printed or electronic form. The author would appreciate notification of typos and other suggestions for improve- ment. ii This is page iii Printer: Opaque this Preface The goal of this book is to introduce statisticians, and other researchers with a background in mathematical statistics, to empirical processes and semiparametric inference. These powerful research techniques are surpris- ingly useful for studying large sample properties of statistical estimates from realistically complex models as well as for developing new and im- proved approaches to statistical inference. This book is more a textbook than a research monograph, although some new results are presented in later chapters. The level of the book is more introductory than the seminal work of van der Vaart and Wellner (1996). In fact, another purpose of this work is to help readers prepare for the mathematically advanced van der Vaart and Wellner text, as well as for the semiparametric inference work of Bickel, Klaassen, Ritov and Wellner (1997). These two books, along with Pollard (1990) and chapters 19 and 25 of van der Vaart (1998), formulate a very complete and successful elucida- tion of modern empirical process methods. The present book owes much by the way of inspiration, concept, and notation to these previous works. What is perhaps new is the introductory, gradual and unified way this book introduces the reader to the field.
    [Show full text]
  • Variable Selection in Semiparametric Regression Modeling
    The Annals of Statistics 2008, Vol. 36, No. 1, 261–286 DOI: 10.1214/009053607000000604 c Institute of Mathematical Statistics, 2008 VARIABLE SELECTION IN SEMIPARAMETRIC REGRESSION MODELING By Runze Li1 and Hua Liang2 Pennsylvania State University and University of Rochester In this paper, we are concerned with how to select significant vari- ables in semiparametric modeling. Variable selection for semipara- metric regression models consists of two components: model selection for nonparametric components and selection of significant variables for the parametric portion. Thus, semiparametric variable selection is much more challenging than parametric variable selection (e.g., linear and generalized linear models) because traditional variable selection procedures including stepwise regression and the best subset selection now require separate model selection for the nonparametric compo- nents for each submodel. This leads to a very heavy computational burden. In this paper, we propose a class of variable selection proce- dures for semiparametric regression models using nonconcave penal- ized likelihood. We establish the rate of convergence of the resulting estimate. With proper choices of penalty functions and regulariza- tion parameters, we show the asymptotic normality of the resulting estimate and further demonstrate that the proposed procedures per- form as well as an oracle procedure. A semiparametric generalized likelihood ratio test is proposed to select significant variables in the nonparametric component. We investigate the asymptotic behavior of the proposed test and demonstrate that its limiting null distri- bution follows a chi-square distribution which is independent of the nuisance parameters. Extensive Monte Carlo simulation studies are conducted to examine the finite sample performance of the proposed variable selection procedures.
    [Show full text]
  • An Averaging Estimator for Two Step M Estimation in Semiparametric Models
    An Averaging Estimator for Two Step M Estimation in Semiparametric Models Ruoyao Shi∗ February, 2021 Abstract In a two step extremum estimation (M estimation) framework with a finite dimen- sional parameter of interest and a potentially infinite dimensional first step nuisance parameter, I propose an averaging estimator that combines a semiparametric esti- mator based on nonparametric first step and a parametric estimator which imposes parametric restrictions on the first step. The averaging weight is the sample analog of an infeasible optimal weight that minimizes the asymptotic quadratic risk. I show that under mild conditions, the asymptotic lower bound of the truncated quadratic risk dif- ference between the averaging estimator and the semiparametric estimator is strictly less than zero for a class of data generating processes (DGPs) that includes both cor- rect specification and varied degrees of misspecification of the parametric restrictions, and the asymptotic upper bound is weakly less than zero. Keywords: two step M estimation, semiparametric model, averaging estimator, uniform dominance, asymptotic quadratic risk JEL Codes: C13, C14, C51, C52 ∗Department of Economics, University of California Riverside, [email protected]. The author thanks Colin Cameron, Xu Cheng, Denis Chetverikov, Yanqin Fan, Jinyong Hahn, Bo Honoré, Toru Kitagawa, Zhipeng Liao, Hyungsik Roger Moon, Whitney Newey, Geert Ridder, Aman Ullah, Haiqing Xu and the participants at various seminars and conferences for helpful comments. This project is generously supported by UC Riverside Regents’ Faculty Fellowship 2019-2020. Zhuozhen Zhao provides great research assistance. All remaining errors are the author’s. 1 1 Introduction Semiparametric models, consisting of a parametric component and a nonparametric com- ponent, have gained popularity in economics.
    [Show full text]
  • Robust Estimates in Generalized Partially Linear Models (With Full Appendix)
    The Annals of Statistics 2006, Vol. 34, No. 6, 2856–2878 DOI: 10.1214/009053606000000858 c Institute of Mathematical Statistics, 2006 ROBUST ESTIMATES IN GENERALIZED PARTIALLY LINEAR MODELS1 By Graciela Boente, Xuming He and Jianhui Zhou Universidad de Buenos Aires and CONICET, University of Illinois at Urbana-Champaign and University of Virginia In this paper, we introduce a family of robust estimates for the parametric and nonparametric components under a generalized par- tially linear model, where the data are modeled by yi|(xi,ti) ∼ F (·, µi) T with µi = H(η(ti)+ xi β), for some known distribution function F and link function H. It is shown that the estimates of β are root-n consistent and asymptotically normal. Through a Monte Carlo study, the performance of these estimators is compared with that of the clas- sical ones. 1. Introduction. Semiparametric models contain both a parametric and a nonparametric component. Sometimes, the nonparametric component plays the role of a nuisance parameter. Much research has been done on estima- tors of the parametric component in a general framework, aiming to obtain asymptotically efficient estimators. The aim of this paper is to consider semi- parametric versions of the generalized linear models where the response y is to be predicted by covariates (x,t), where x Rp and t R. It will be assumed that the conditional distribution of y∈(x,t) belongs∈T to ⊂ the canonical exponential family exp[yθ(x,t) B(θ(x,t)) +|C(y)] for known functions B and C. Then µ(x,t)= E(y (x,t))− = B′(θ(x,t)), with B′ denoting the deriva- tive of B.
    [Show full text]
  • Semiparametric Theory, Using As a Running Example the Common Problem of Estimating an Average Causal Effect
    Semiparametric theory Edward H. Kennedy∗ Carnegie Mellon University, PA, USA September 20, 2017 Abstract In this paper we give a brief review of semiparametric theory, using as a running example the common problem of estimating an average causal effect. Semiparamet- ric models allow at least part of the data-generating process to be unspecified and unrestricted, and can often yield robust estimators that nonetheless behave similarly to those based on parametric likelihood assumptions, e.g., fast rates of convergence to normal limiting distributions. We discuss the basics of semiparametric theory, focusing on influence functions. Keywords: causal inference, double robustness, efficiency theory, functional estimation, influence functions, nonparametric theory, robust inference, tangent space. arXiv:1709.06418v1 [stat.ME] 15 Sep 2017 ∗Edward Kennedy ([email protected]) is Assistant Professor in the Department of Statistics, Carnegie Mellon University, Pittsburgh, PA 15213. In this paper we give a review of semiparametric theory, using as a running example the common problem of estimating an average causal effect. Our review draws heavily on foundational work in general semiparametrics by Begun et al. (1983), Bickel et al. (1993), Pfanzagl (1982), van der Vaart (2000), and van der Vaart (2002), among others (Newey 1994; Kosorok 2007), as well as many modern developments in missing data and causal inference problems by Robins & van der Laan (Robins 1986; Robins et al. 1994; Robins et al. 1995; Robins 2000; van der Laan & Robins 2003; van der Laan & Rose 2011; van der Vaart 2014; Robins et al. 2017), and colleagues (Hahn 1998; Tsiatis 2006). We refer to Tsiatis (2006) for a very readable review with more details.
    [Show full text]
  • Semiparametric Regression in Stata
    Introduction PLM Stata Semipar Heteroskedasticity Endogeneity Heterogeneity Mfx Single index Semiparametric regression in Stata Vincenzo Verardi 2013 UK Stata Users Group meeting London, UK September 2013 Vincenzo Verardi Semiparametric regression 12/09/2013 1/66 Introduction PLM Stata Semipar Heteroskedasticity Endogeneity Heterogeneity Mfx Single index Introduction Semiparametric regression models Semiparametric regression Deals with the introduction of some very general non-linear functional forms in regression analysis Generally used to …t a parametric model in which the functional form of a subset of the explanatory variables is not known and/or in which the distribution of the error term cannot be assumed to be of a speci…c type beforehand. Most popular semiparametric regression models are the partially linear models and single index models Remark: for an excellent description of semiparametric estimators, we advice to read the book by Ahamada and Flachaire (2010) from which many of the explanations available here come from. Vincenzo Verardi Semiparametric regression 12/09/2013 2/66 Introduction PLM Stata Semipar Heteroskedasticity Endogeneity Heterogeneity Mfx Single index Introduction Semiparametric regression models Partially linear models The partially linear model is de…ned as: y = X β + m(z) + ε Advantage 1: This model allows "any" form of the unknown function m Advantage 2: βˆ is pn-consistent Single index models The single index model is de…ned as: y = g(X β) + ε Advantage 1: generalizes the linear regression model (which assumes g( ) is linear) Advantage 2: the curse of dimensionality is avoided as there is only one nonparametric dimension Vincenzo Verardi Semiparametric regression 12/09/2013 3/66 Introduction PLM Stata Semipar Heteroskedasticity Endogeneity Heterogeneity Mfx Single index Introduction PLM example Hedonic pricing equation of houses Wooldridge (2000): What was the e¤ect of a local garbage incinerator on housing prices in North Andover in 1981? .
    [Show full text]
  • Semiparametric Regression Pursuit
    Statistica Sinica 22 (2012), 1403-1426 doi:http://dx.doi.org/10.5705/ss.2010.298 SEMIPARAMETRIC REGRESSION PURSUIT Jian Huang, Fengrong Wei and Shuangge Ma University of Iowa, University of West Georgia and Yale University Abstract: The semiparametric partially linear model allows flexible modeling of covariate effects on the response variable in regression. It combines the flexibility of nonparametric regression and the parsimony of linear regression. The most important assumption in the existing methods for the estimation in this model is that a priori it is known which covariates have a linear effect and which do not. However, in applied work, this is rarely known in advance. We consider the problem of estimation in the partially linear models without assuming a priori which covariates have linear effects. We propose a semiparametric regression pursuit method for identifying the covariates with a linear effect. Our proposed method is a penalized regression approach using a group minimax concave penalty. Under suitable conditions we show that the proposed approach is model-pursuit consistent, meaning that it can correctly determine which covariates have a linear effect and which do not with high probability. The performance of the proposed method is evaluated using simulation studies that support our theoretical results. A data example is used to illustrated the application of the proposed method. Key words and phrases: Group selection, minimax concave penalty, model-pursuit consistency, penalized regression, semiparametric models, structure estimation. 1. Introduction Suppose we have a random sample (yi; xi1; : : : ; xip); 1 ≤ i ≤ n, where yi is the response variable and (xi1; : : : ; xip) is a p-dimensional covariate vector.
    [Show full text]
  • Parametric Versus Semi and Nonparametric Regression Models
    Parametric versus Semi and Nonparametric Regression Models Hamdy F. F. Mahmoud1;2 1 Department of Statistics, Virginia Tech, Blacksburg VA, USA. 2 Department of Statistics, Mathematics and Insurance, Assiut University, Egypt. Abstract Three types of regression models researchers need to be familiar with and know the requirements of each: parametric, semiparametric and nonparametric regression models. The type of modeling used is based on how much information are available about the form of the relationship between response variable and explanatory variables, and the random error distribution. In this article, differences between models, common methods of estimation, robust estimation, and applications are introduced. The R code for all the graphs and analyses presented here, in this article, is available in the Appendix. Keywords: Parametric, semiparametric, nonparametric, single index model, robust estimation, kernel regression, spline smoothing. 1 Introduction The aim of this paper is to answer many questions researchers may have when they fit their data, such as what are the differences between parametric, semi/nonparamertic models? which one should be used to model a real data set? what estimation method should be used?, and which type of modeling is better? These questions and others are addressed by examples. arXiv:1906.10221v1 [stat.ME] 24 Jun 2019 Assume that a researcher collected data about Y, a response variable, and explanatory variables, X = (x1; x2; : : : ; xk). The model that describes the relationship between Y and X can be written as Y = f(X; β) + , (1) where β is a vector of k parameters, is an error term whose distribution may or may not be normal, f(·) is some function that describe the relationship between Y and X.
    [Show full text]