Generalized Linear Models

Total Page:16

File Type:pdf, Size:1020Kb

Generalized Linear Models Chapter 6 Generalized Linear Models In Chapters 2 and 4 we studied how to estimate simple probability densities over a single random variable—that is, densities of the form P (Y ). In this chapter we move on to the problem of estimating conditional densities—that is, densities of the form P (Y X). Logically speaking, it would be possible to deal with this problem simply by assuming| that Y may have an arbitrarily different distribution for each possible value of X, and to use techniques we’ve covered already to estimate a different density P (Y X = xi) for each possible value xi of X. However, this approach would miss the fact that X may| have a systematic effect on Y ; missing this fact when it is true would make it much more difficult to estimate the conditional distribution. Here, we cover a popular family of conditional probability distributions known as generalized linear models. These models can be used for a wide range of data types and have attractive computational properties. 6.1 The form of the generalized linear model Suppose we are interested in modeling the distribution of Y conditional on a number of random variables X1,...,Xn. Generalized linear models are a framework for modeling this type of conditional distribution P (Y X ,...,X ) subject to four key assumptions: | 1 n 1. The influences of the Xi variables on Y can be summarized into an intermediate form, the linear predictor{ } η; 2. η is a linear combination of the X ; { i} 3. There is a smooth, invertible function l mapping η to the expected value µ of Y ; 4. The distribution P (Y = y; µ) of Y around µ is a member of a certain class of noise 1 functions and is not otherwise sensitive to the Xi variables. Assumptions 1 through 3 can be expressed by the following two equations: 1The class of allowable noise functions is described in Section ??. 107 θ X η Y Figure 6.1: A graphical depiction of the generalized linear model. The influence of the conditioning variables X on the response Y is completely mediated by the linear predictor η. η = α + β X + + β X (linearpredictor) (6.1) 1 1 ··· n n η =l(µ) (linkfunction) (6.2) Assumption 4 implies conditional independence of Y from the Xi variables given η. Various choices of l(µ) and P (Y = y; µ) give us different classes{ } of models appropriate for different types of data. In all cases, we can estimate the parameters of the models using any of likelihood-based techniques discussed in Chapter 4. We cover three common classes of such models in this chapter: linear models, logit (logistic) models, and log-linear models. 6.2 Linear models and linear regression We can obtain the classic linear model by chooosing the identity link function η = l(µ)= µ and a noise function that adds noise ǫ N(0,σ2) ∼ to the mean µ. Substituting these in to Equations (6.1) and 6.2, we can write Y directly as a function of X as follows: { i} Predicted Mean Noise N(0,σ2) ∼ Y = α + β X + + β X + ǫ (6.3) 1 1 ··· n n z }| { z}|{ 2 We can also write this whole thing in more compressed form as Y N(α i βiXi,σ ). To gain intuitions for how this model places a conditional probability∼ density on Y , we can visualize this probability density for a single independent variable X, asP in Figure 6.2— lighter means more probable. Each vertical slice of constant X = x represents a conditional Roger Levy – Probabilistic Models in the Study of Language draft, November 6, 2012 108 Conditional probability density P(y|x) for a linear model 4 0.20 4 2 0.15 2 y 0 y 0 0.10 −2 −2 0.05 −4 −4 0.00 −4 −2 0 2 4 −4 −2 0 2 4 x x (a) Predicted outcome (b) Probability density Figure 6.2: A plot of the probability density on the outcome of the Y random variable given 1 2 the X random variable; in this case we have η = 2 X and σ = 4. distribution P (Y x). If you imagine a vertical line extending through the plot at X = 0, you | will see that the plot along this line is lightest in color at Y = 1 = α. This is the point at which ǫ takes its most probable value, 0. For this reason, α is− also called the intercept parameter of the model, and the βi are called the slope parameters. 6.2.1 Fitting a linear model The process of estimating the parameters α and βi of a linear model on the basis of some data (also called fitting the model to the data) is called linear regression. There are many techniques for parameter estimation in linear regression; here we will cover the method of maximum likelihood and also Bayesian linear regression. Maximum-likelihood linear regression Before we talk about exactly what the maximum-likelihood estimate looks like, we’ll intro- duce some useful terminology. Suppose that we have chosen model parametersα ˆ and βˆi . This means that for each point x ,y in our dataset y, we can construct a predicted{ } h j ji value fory ˆj as follows: yˆj =α ˆ + βˆ1xj1 + ... βˆnxjn where xji is the value of the i-th predictor variable for the j-th data point. This predicted value is both the expectated value and the modal value of Yj due to the Gaussian-noise assumption of linear regression. We define the residual of the j-th data point simply as Roger Levy – Probabilistic Models in the Study of Language draft, November 6, 2012 109 y yˆ j − j —that is, the amount by which our model’s prediction missed the observed value. It turns out that for linear models with a normally-distributed error term ǫ, the log- likelihood of the model parameters with respect to y is proportional to the sum of the squared residuals. This means that the maximum-likelihood estimate of the parameters is also the estimate that minimizes the the sum of the squared residuals. You will often see description of regression models being fit using least-squares estimation. Whenever you see this, recognize that this is equivalent to maximum-likelihood estimation under the assumption that residual error is normally-distributed. 6.2.2 Fitting a linear model: case study The dataset english contains reaction times for lexical decision and naming of isolated English words, as well as written frequencies for those words. Reaction times are measured in milliseconds, and word frequencies are measured in appearances in a 17.9-million word written corpus. (All these variables are recorded in log-space) It is well-established that words of high textual frequency are generally responded to more quickly than words of low textual frequency. Let us consider a linear model in which reaction time RT depends on the log-frequency, F , of the word: RT = α + βF F + ǫ (6.4) This linear model corresponds to a formula in R, which can be specified in either of the Introduction to following ways: section 11.1 RT ~ F RT~1+F The 1 in the latter formula refers to the intercept of the model; the presence of an intercept is implicit in the first formula. The result of the linear regression is an intercept α = 843.58 and a slope β = 29.76. F − The WrittenFrequency variable is in natural log-space, so the slope can be interpreted as saying that if two words differ in frequency by a factor of e 2.718, then on average the more frequent word will be recognized as a word of English 26.97≈ milliseconds faster than the less frequent word. The intercept, 843.58, is the predicted reaction time for a word whose log-frequency is 0—that is, a word occurring only once in the corpus. 6.2.3 Conceptual underpinnings of best linear fit Let us now break down how the model goes about fitting data in a simple example. Suppose we have only three observations of log-frequency/RT pairs: Roger Levy – Probabilistic Models in the Study of Language draft, November 6, 2012 110 exp(RTlexdec) 600 800 1000 1200 0 2 4 6 8 10 WrittenFrequency Figure 6.3: Lexical decision reaction times as a function of word frequency 4, 800 h i 6, 775 h i 8, 700 h i Let use consider four possible parameter estimates for these data points. Three estimates will draw a line through two of the points and miss the third; the last estimate will draw a line that misses but is reasonably close to all the points. First consider the solid black line, which has intercept 910 and slope -25. It predicts the following values, missing all three points: x yˆ Residual (ˆy y) 4 810 10 − 6760− 15 8 710 10 − and the sum of its squared residuals is 425. Each of the other three lines has only one non-zero residual, but that residual is much larger, and in all three case, the sum of squared residuals is larger than for the solid black line. This means that the likelihood of the parameter values α = 910, βF = 25 is higher than the likelihood of the parameters corresponding to any of the other lines.− What is the MLE for α, βF with respect to these three data points, and what are the residuals for the MLE? Results of a linear fit (almost every statistical software package supports linear regression) indicate that the MLE is α = 908 1 , β = 25.
Recommended publications
  • Generalized Linear Models and Generalized Additive Models
    00:34 Friday 27th February, 2015 Copyright ©Cosma Rohilla Shalizi; do not distribute without permission updates at http://www.stat.cmu.edu/~cshalizi/ADAfaEPoV/ Chapter 13 Generalized Linear Models and Generalized Additive Models [[TODO: Merge GLM/GAM and Logistic Regression chap- 13.1 Generalized Linear Models and Iterative Least Squares ters]] [[ATTN: Keep as separate Logistic regression is a particular instance of a broader kind of model, called a gener- chapter, or merge with logis- alized linear model (GLM). You are familiar, of course, from your regression class tic?]] with the idea of transforming the response variable, what we’ve been calling Y , and then predicting the transformed variable from X . This was not what we did in logis- tic regression. Rather, we transformed the conditional expected value, and made that a linear function of X . This seems odd, because it is odd, but it turns out to be useful. Let’s be specific. Our usual focus in regression modeling has been the condi- tional expectation function, r (x)=E[Y X = x]. In plain linear regression, we try | to approximate r (x) by β0 + x β. In logistic regression, r (x)=E[Y X = x] = · | Pr(Y = 1 X = x), and it is a transformation of r (x) which is linear. The usual nota- tion says | ⌘(x)=β0 + x β (13.1) · r (x) ⌘(x)=log (13.2) 1 r (x) − = g(r (x)) (13.3) defining the logistic link function by g(m)=log m/(1 m). The function ⌘(x) is called the linear predictor. − Now, the first impulse for estimating this model would be to apply the transfor- mation g to the response.
    [Show full text]
  • A Generalized Linear Model for Principal Component Analysis of Binary Data
    A Generalized Linear Model for Principal Component Analysis of Binary Data Andrew I. Schein Lawrence K. Saul Lyle H. Ungar Department of Computer and Information Science University of Pennsylvania Moore School Building 200 South 33rd Street Philadelphia, PA 19104-6389 {ais,lsaul,ungar}@cis.upenn.edu Abstract they are not generally appropriate for other data types. Recently, Collins et al.[5] derived generalized criteria We investigate a generalized linear model for for dimensionality reduction by appealing to proper- dimensionality reduction of binary data. The ties of distributions in the exponential family. In their model is related to principal component anal- framework, the conventional PCA of real-valued data ysis (PCA) in the same way that logistic re- emerges naturally from assuming a Gaussian distribu- gression is related to linear regression. Thus tion over a set of observations, while generalized ver- we refer to the model as logistic PCA. In this sions of PCA for binary and nonnegative data emerge paper, we derive an alternating least squares respectively by substituting the Bernoulli and Pois- method to estimate the basis vectors and gen- son distributions for the Gaussian. For binary data, eralized linear coefficients of the logistic PCA the generalized model's relationship to PCA is anal- model. The resulting updates have a simple ogous to the relationship between logistic and linear closed form and are guaranteed at each iter- regression[12]. In particular, the model exploits the ation to improve the model's likelihood. We log-odds as the natural parameter of the Bernoulli dis- evaluate the performance of logistic PCA|as tribution and the logistic function as its canonical link.
    [Show full text]
  • Generalized Linear Models (Glms)
    San Jos´eState University Math 261A: Regression Theory & Methods Generalized Linear Models (GLMs) Dr. Guangliang Chen This lecture is based on the following textbook sections: • Chapter 13: 13.1 – 13.3 Outline of this presentation: • What is a GLM? • Logistic regression • Poisson regression Generalized Linear Models (GLMs) What is a GLM? In ordinary linear regression, we assume that the response is a linear function of the regressors plus Gaussian noise: 0 2 y = β0 + β1x1 + ··· + βkxk + ∼ N(x β, σ ) | {z } |{z} linear form x0β N(0,σ2) noise The model can be reformulate in terms of • distribution of the response: y | x ∼ N(µ, σ2), and • dependence of the mean on the predictors: µ = E(y | x) = x0β Dr. Guangliang Chen | Mathematics & Statistics, San Jos´e State University3/24 Generalized Linear Models (GLMs) beta=(1,2) 5 4 3 β0 + β1x b y 2 y 1 0 −1 0.0 0.2 0.4 0.6 0.8 1.0 x x Dr. Guangliang Chen | Mathematics & Statistics, San Jos´e State University4/24 Generalized Linear Models (GLMs) Generalized linear models (GLM) extend linear regression by allowing the response variable to have • a general distribution (with mean µ = E(y | x)) and • a mean that depends on the predictors through a link function g: That is, g(µ) = β0x or equivalently, µ = g−1(β0x) Dr. Guangliang Chen | Mathematics & Statistics, San Jos´e State University5/24 Generalized Linear Models (GLMs) In GLM, the response is typically assumed to have a distribution in the exponential family, which is a large class of probability distributions that have pdfs of the form f(x | θ) = a(x)b(θ) exp(c(θ) · T (x)), including • Normal - ordinary linear regression • Bernoulli - Logistic regression, modeling binary data • Binomial - Multinomial logistic regression, modeling general cate- gorical data • Poisson - Poisson regression, modeling count data • Exponential, Gamma - survival analysis Dr.
    [Show full text]
  • Regression Summary Project Analysis for Today Review Questions: Categorical Variables
    Statistics 102 Regression Summary Spring, 2000 - 1 - Regression Summary Project Analysis for Today First multiple regression • Interpreting the location and wiring coefficient estimates • Interpreting interaction terms • Measuring significance Second multiple regression • Deciding how to extend a model • Diagnostics (leverage plots, residual plots) Review Questions: Categorical Variables Where do those terms in a categorical regression come from? • You cannot use a categorical term directly in regression (e.g. 2(“Yes”)=?). • JMP converts each categorical variable into a collection of numerical variables that represent the information in the categorical variable, but are numerical and so can be used in regression. • These special variables (a.k.a., dummy variables) use only the numbers +1, 0, and –1. • A categorical variable with k categories requires (k-1) of these special numerical variables. Thus, adding a categorical variable with, for example, 5 categories adds 4 of these numerical variables to the model. How do I use the various tests in regression? What question are you trying to answer… • Does this predictor add significantly to my model, improving the fit beyond that obtained with the other predictors? (t-ratio, CI, p-value) • Does this collection of predictors add significantly to my model? Partial-F (Note: the only time you need partial F is when working with categorical variables that define 3 or more categories. In these cases, JMP shows you the partial-F as an “Effect Test”.) • Does my full model explain “more than random variation”? Use the F-ratio from the Anova summary table. Statistics 102 Regression Summary Spring, 2000 - 2 - How do I interpret JMP output with categorical variables? Term Estimate Std Error t Ratio Prob>|t| Intercept 179.59 5.62 32.0 0.00 Run Size 0.23 0.02 9.5 0.00 Manager[a-c] 22.94 7.76 3.0 0.00 Manager[b-c] 6.90 8.73 0.8 0.43 Manager[a-c]*Run Size 0.07 0.04 2.1 0.04 Manager[b-c]*Run Size -0.10 0.04 -2.6 0.01 • Brackets denote the JMP’s version of dummy variables.
    [Show full text]
  • Logistic Regression, Dependencies, Non-Linear Data and Model Reduction
    COMP6237 – Logistic Regression, Dependencies, Non-linear Data and Model Reduction Markus Brede [email protected] Lecture slides available here: http://users.ecs.soton.ac.uk/mb8/stats/datamining.html (Thanks to Jason Noble and Cosma Shalizi whose lecture materials I used to prepare) COMP6237: Logistic Regression ● Outline: – Introduction – Basic ideas of logistic regression – Logistic regression using R – Some underlying maths and MLE – The multinomial case – How to deal with non-linear data ● Model reduction and AIC – How to deal with dependent data – Summary – Problems Introduction ● Previous lecture: Linear regression – tried to predict a continuous variable from variation in another continuous variable (E.g. basketball ability from height) ● Here: Logistic regression – Try to predict results of a binary (or categorical) outcome variable Y from a predictor variable X – This is a classification problem: classify X as belonging to one of two classes – Occurs quite often in science … e.g. medical trials (will a patient live or die dependent on medication?) Dependent variable Y Predictor Variables X The Oscars Example ● A fictional data set that looks at what it takes for a movie to win an Oscar ● Outcome variable: Oscar win, yes or no? ● Predictor variables: – Box office takings in millions of dollars – Budget in millions of dollars – Country of origin: US, UK, Europe, India, other – Critical reception (scores 0 … 100) – Length of film in minutes – This (fictitious) data set is available here: https://www.southampton.ac.uk/~mb1a10/stats/filmData.txt Predicting Oscar Success ● Let's start simple and look at only one of the predictor variables ● Do big box office takings make Oscar success more likely? ● Could use same techniques as below to look at budget size, film length, etc.
    [Show full text]
  • Bayesian Inference Chapter 4: Regression and Hierarchical Models
    Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Aus´ınand Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering Conchi Aus´ınand Mike Wiper Regression and hierarchical models Masters Programmes 1 / 35 Objective AFM Smith Dennis Lindley We analyze the Bayesian approach to fitting normal and generalized linear models and introduce the Bayesian hierarchical modeling approach. Also, we study the modeling and forecasting of time series. Conchi Aus´ınand Mike Wiper Regression and hierarchical models Masters Programmes 2 / 35 Contents 1 Normal linear models 1.1. ANOVA model 1.2. Simple linear regression model 2 Generalized linear models 3 Hierarchical models 4 Dynamic models Conchi Aus´ınand Mike Wiper Regression and hierarchical models Masters Programmes 3 / 35 Normal linear models A normal linear model is of the following form: y = Xθ + ; 0 where y = (y1;:::; yn) is the observed data, X is a known n × k matrix, called 0 the design matrix, θ = (θ1; : : : ; θk ) is the parameter set and follows a multivariate normal distribution. Usually, it is assumed that: 1 ∼ N 0 ; I : k φ k A simple example of normal linear model is the simple linear regression model T 1 1 ::: 1 where X = and θ = (α; β)T . x1 x2 ::: xn Conchi Aus´ınand Mike Wiper Regression and hierarchical models Masters Programmes 4 / 35 Normal linear models Consider a normal linear model, y = Xθ + . A conjugate prior distribution is a normal-gamma distribution:
    [Show full text]
  • How Prediction Statistics Can Help Us Cope When We Are Shaken, Scared and Irrational
    EGU21-15219 https://doi.org/10.5194/egusphere-egu21-15219 EGU General Assembly 2021 © Author(s) 2021. This work is distributed under the Creative Commons Attribution 4.0 License. How prediction statistics can help us cope when we are shaken, scared and irrational Yavor Kamer1, Shyam Nandan2, Stefan Hiemer1, Guy Ouillon3, and Didier Sornette4,5 1RichterX.com, Zürich, Switzerland 2Windeggstrasse 5, 8953 Dietikon, Zurich, Switzerland 3Lithophyse, Nice, France 4Department of Management,Technology and Economics, ETH Zürich, Zürich, Switzerland 5Institute of Risk Analysis, Prediction and Management (Risks-X), SUSTech, Shenzhen, China Nature is scary. You can be sitting at your home and next thing you know you are trapped under the ruble of your own house or sucked into a sinkhole. For millions of years we have been the figurines of this precarious scene and we have found our own ways of dealing with the anxiety. It is natural that we create and consume prophecies, conspiracies and false predictions. Information technologies amplify not only our rational but also irrational deeds. Social media algorithms, tuned to maximize attention, make sure that misinformation spreads much faster than its counterpart. What can we do to minimize the adverse effects of misinformation, especially in the case of earthquakes? One option could be to designate one authoritative institute, set up a big surveillance network and cancel or ban every source of misinformation before it spreads. This might have worked a few centuries ago but not in this day and age. Instead we propose a more inclusive option: embrace all voices and channel them into an actual, prospective earthquake prediction platform (Kamer et al.
    [Show full text]
  • Generalized and Weighted Least Squares Estimation
    LINEAR REGRESSION ANALYSIS MODULE – VII Lecture – 25 Generalized and Weighted Least Squares Estimation Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur 2 The usual linear regression model assumes that all the random error components are identically and independently distributed with constant variance. When this assumption is violated, then ordinary least squares estimator of regression coefficient looses its property of minimum variance in the class of linear and unbiased estimators. The violation of such assumption can arise in anyone of the following situations: 1. The variance of random error components is not constant. 2. The random error components are not independent. 3. The random error components do not have constant variance as well as they are not independent. In such cases, the covariance matrix of random error components does not remain in the form of an identity matrix but can be considered as any positive definite matrix. Under such assumption, the OLSE does not remain efficient as in the case of identity covariance matrix. The generalized or weighted least squares method is used in such situations to estimate the parameters of the model. In this method, the deviation between the observed and expected values of yi is multiplied by a weight ω i where ω i is chosen to be inversely proportional to the variance of yi. n 2 For simple linear regression model, the weighted least squares function is S(,)ββ01=∑ ωii( yx −− β0 β 1i) . ββ The least squares normal equations are obtained by differentiating S (,) ββ 01 with respect to 01 and and equating them to zero as nn n ˆˆ β01∑∑∑ ωβi+= ωiixy ω ii ii=11= i= 1 n nn ˆˆ2 βω01∑iix+= βω ∑∑ii x ω iii xy.
    [Show full text]
  • 18.650 (F16) Lecture 10: Generalized Linear Models (Glms)
    Statistics for Applications Chapter 10: Generalized Linear Models (GLMs) 1/52 Linear model A linear model assumes Y X (µ(X), σ2I), | ∼ N And ⊤ IE(Y X) = µ(X) = X β, | 2/52 Components of a linear model The two components (that we are going to relax) are 1. Random component: the response variable Y X is continuous | and normally distributed with mean µ = µ(X) = IE(Y X). | 2. Link: between the random and covariates X = (X(1),X(2), ,X(p))⊤: µ(X) = X⊤β. · · · 3/52 Generalization A generalized linear model (GLM) generalizes normal linear regression models in the following directions. 1. Random component: Y some exponential family distribution ∼ 2. Link: between the random and covariates: ⊤ g µ(X) = X β where g called link function� and� µ = IE(Y X). | 4/52 Example 1: Disease Occuring Rate In the early stages of a disease epidemic, the rate at which new cases occur can often increase exponentially through time. Hence, if µi is the expected number of new cases on day ti, a model of the form µi = γ exp(δti) seems appropriate. ◮ Such a model can be turned into GLM form, by using a log link so that log(µi) = log(γ) + δti = β0 + β1ti. ◮ Since this is a count, the Poisson distribution (with expected value µi) is probably a reasonable distribution to try. 5/52 Example 2: Prey Capture Rate(1) The rate of capture of preys, yi, by a hunting animal, tends to increase with increasing density of prey, xi, but to eventually level off, when the predator is catching as much as it can cope with.
    [Show full text]
  • Using Survey Data Author: Jen Buckley and Sarah King-Hele Updated: August 2015 Version: 1
    ukdataservice.ac.uk Using survey data Author: Jen Buckley and Sarah King-Hele Updated: August 2015 Version: 1 Acknowledgement/Citation These pages are based on the following workbook, funded by the Economic and Social Research Council (ESRC). Williamson, Lee, Mark Brown, Jo Wathan, Vanessa Higgins (2013) Secondary Analysis for Social Scientists; Analysing the fear of crime using the British Crime Survey. Updated version by Sarah King-Hele. Centre for Census and Survey Research We are happy for our materials to be used and copied but request that users should: • link to our original materials instead of re-mounting our materials on your website • cite this as an original source as follows: Buckley, Jen and Sarah King-Hele (2015). Using survey data. UK Data Service, University of Essex and University of Manchester. UK Data Service – Using survey data Contents 1. Introduction 3 2. Before you start 4 2.1. Research topic and questions 4 2.2. Survey data and secondary analysis 5 2.3. Concepts and measurement 6 2.4. Change over time 8 2.5. Worksheets 9 3. Find data 10 3.1. Survey microdata 10 3.2. UK Data Service 12 3.3. Other ways to find data 14 3.4. Evaluating data 15 3.5. Tables and reports 17 3.6. Worksheets 18 4. Get started with survey data 19 4.1. Registration and access conditions 19 4.2. Download 20 4.3. Statistics packages 21 4.4. Survey weights 22 4.5. Worksheets 24 5. Data analysis 25 5.1. Types of variables 25 5.2. Variable distributions 27 5.3.
    [Show full text]
  • Linear, Ridge Regression, and Principal Component Analysis
    Linear, Ridge Regression, and Principal Component Analysis Linear, Ridge Regression, and Principal Component Analysis Jia Li Department of Statistics The Pennsylvania State University Email: [email protected] http://www.stat.psu.edu/∼jiali Jia Li http://www.stat.psu.edu/∼jiali Linear, Ridge Regression, and Principal Component Analysis Introduction to Regression I Input vector: X = (X1, X2, ..., Xp). I Output Y is real-valued. I Predict Y from X by f (X ) so that the expected loss function E(L(Y , f (X ))) is minimized. I Square loss: L(Y , f (X )) = (Y − f (X ))2 . I The optimal predictor ∗ 2 f (X ) = argminf (X )E(Y − f (X )) = E(Y | X ) . I The function E(Y | X ) is the regression function. Jia Li http://www.stat.psu.edu/∼jiali Linear, Ridge Regression, and Principal Component Analysis Example The number of active physicians in a Standard Metropolitan Statistical Area (SMSA), denoted by Y , is expected to be related to total population (X1, measured in thousands), land area (X2, measured in square miles), and total personal income (X3, measured in millions of dollars). Data are collected for 141 SMSAs, as shown in the following table. i : 1 2 3 ... 139 140 141 X1 9387 7031 7017 ... 233 232 231 X2 1348 4069 3719 ... 1011 813 654 X3 72100 52737 54542 ... 1337 1589 1148 Y 25627 15389 13326 ... 264 371 140 Goal: Predict Y from X1, X2, and X3. Jia Li http://www.stat.psu.edu/∼jiali Linear, Ridge Regression, and Principal Component Analysis Linear Methods I The linear regression model p X f (X ) = β0 + Xj βj .
    [Show full text]
  • Simple Linear Regression with Least Square Estimation: an Overview
    Aditya N More et al, / (IJCSIT) International Journal of Computer Science and Information Technologies, Vol. 7 (6) , 2016, 2394-2396 Simple Linear Regression with Least Square Estimation: An Overview Aditya N More#1, Puneet S Kohli*2, Kshitija H Kulkarni#3 #1-2Information Technology Department,#3 Electronics and Communication Department College of Engineering Pune Shivajinagar, Pune – 411005, Maharashtra, India Abstract— Linear Regression involves modelling a relationship amongst dependent and independent variables in the form of a (2.1) linear equation. Least Square Estimation is a method to determine the constants in a Linear model in the most accurate way without much complexity of solving. Metrics where such as Coefficient of Determination and Mean Square Error is the ith value of the sample data point determine how good the estimation is. Statistical Packages is the ith value of y on the predicted regression such as R and Microsoft Excel have built in tools to perform Least Square Estimation over a given data set. line The above equation can be geometrically depicted by Keywords— Linear Regression, Machine Learning, Least Squares Estimation, R programming figure 2.1. If we draw a square at each point whose length is equal to the absolute difference between the sample data point and the predicted value as shown, each of the square would then represent the residual error in placing the I. INTRODUCTION regression line. The aim of the least square method would Linear Regression involves establishing linear be to place the regression line so as to minimize the sum of relationships between dependent and independent variables.
    [Show full text]