Review of “Technical Note: Effects of Uncertainties and Number of Data Points on Line Fitting – a Case Study on New Particle Formation”, Revised, by Mikkonen Et Al
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Least-Squares Fitting of Circles and Ellipses∗
Least-Squares Fitting of Circles and Ellipses∗ Walter Gander Gene H. Golub Rolf Strebel Dedicated to Ake˚ Bj¨orck on the occasion of his 60th birthday. Abstract Fitting circles and ellipses to given points in the plane is a problem that arises in many application areas, e.g. computer graphics [1], coordinate metrol- ogy [2], petroleum engineering [11], statistics [7]. In the past, algorithms have been given which fit circles and ellipses in some least squares sense without minimizing the geometric distance to the given points [1], [6]. In this paper we present several algorithms which compute the ellipse for which the sum of the squares of the distances to the given points is minimal. These algorithms are compared with classical simple and iterative methods. Circles and ellipses may be represented algebraically i.e. by an equation of the form F(x) = 0. If a point is on the curve then its coordinates x are a zero of the function F. Alternatively, curves may be represented in parametric form, which is well suited for minimizing the sum of the squares of the distances. 1 Preliminaries and Introduction Ellipses, for which the sum of the squares of the distances to the given points is minimal will be referred to as \best fit” or \geometric fit", and the algorithms will be called \geometric". Determining the parameters of the algebraic equation F (x)=0intheleast squares sense will be denoted by \algebraic fit" and the algorithms will be called \algebraic". We will use the well known Gauss-Newton method to solve the nonlinear least T squares problem (cf. -
Ordinary Least Squares 1 Ordinary Least Squares
Ordinary least squares 1 Ordinary least squares In statistics, ordinary least squares (OLS) or linear least squares is a method for estimating the unknown parameters in a linear regression model. This method minimizes the sum of squared vertical distances between the observed responses in the dataset and the responses predicted by the linear approximation. The resulting estimator can be expressed by a simple formula, especially in the case of a single regressor on the right-hand side. The OLS estimator is consistent when the regressors are exogenous and there is no Okun's law in macroeconomics states that in an economy the GDP growth should multicollinearity, and optimal in the class of depend linearly on the changes in the unemployment rate. Here the ordinary least squares method is used to construct the regression line describing this law. linear unbiased estimators when the errors are homoscedastic and serially uncorrelated. Under these conditions, the method of OLS provides minimum-variance mean-unbiased estimation when the errors have finite variances. Under the additional assumption that the errors be normally distributed, OLS is the maximum likelihood estimator. OLS is used in economics (econometrics) and electrical engineering (control theory and signal processing), among many areas of application. Linear model Suppose the data consists of n observations { y , x } . Each observation includes a scalar response y and a i i i vector of predictors (or regressors) x . In a linear regression model the response variable is a linear function of the i regressors: where β is a p×1 vector of unknown parameters; ε 's are unobserved scalar random variables (errors) which account i for the discrepancy between the actually observed responses y and the "predicted outcomes" x′ β; and ′ denotes i i matrix transpose, so that x′ β is the dot product between the vectors x and β. -
Lesson 4 – Instrumental Variables
Ph.D. IN ENGINEERING AND Advanced methods for system APPLIED SCIENCES identification Ph.D. course Lesson 4: Instrumental variables TEACHER Mirko Mazzoleni PLACE University of Bergamo Outline 1. Introduction to error-in-variables problems 2. Least Squares revised 3. The instrumental variable method 4. Estimate an ARX model from ARMAX generated data 5. Application study: the VRFT approach for direct data driven control 6. Conclusions 2 /44 Outline 1. Introduction to error-in-variables problems 2. Least Squares revised 3. The instrumental variable method 4. Estimate an ARX model from ARMAX generated data 5. Application study: the VRFT approach for direct data driven control 6. Conclusions 3 /44 Introduction Many different solutions have been presented for system identification of linear dynamic systems from noise-corrupted output measurements On the other hand, estimation of the parameters for linear dynamic systems when also the input is affected by noise is recognized as a more difficult problem Representations where errors or measurement noises are present on both inputs and outputs are usually called Errors-in-Variables (EIV) models In case of static systems, errors-in-variables representations are closely related to other well-known topics such as latent variables models and factor models 4 /44 Introduction Errors-in-variables models can be motivated in several situations: • modeling of the dynamics between the noise-free input and noise-free output. The reason can be to have a better understanding of the underlying relations (e.g. in econometrics), rather than to make a good prediction from noisy data • when the user lacks enough information to classify the available signals into inputs and outputs and prefer to use a “symmetric” system model. -
Overview of Total Least Squares Methods
Overview of total least squares methods ,1 2 Ivan Markovsky∗ and Sabine Van Huffel 1 — School of Electronics and Computer Science, University of Southampton, SO17 1BJ, UK, [email protected] 2 — Katholieke Universiteit Leuven, ESAT-SCD/SISTA, Kasteelpark Arenberg 10 B–3001 Leuven, Belgium, [email protected] Abstract We review the development and extensions of the classical total least squares method and describe algorithms for its generalization to weighted and structured approximation problems. In the generic case, the classical total least squares problem has a unique solution, which is given in analytic form in terms of the singular value de- composition of the data matrix. The weighted and structured total least squares problems have no such analytic solution and are currently solved numerically by local optimization methods. We explain how special structure of the weight matrix and the data matrix can be exploited for efficient cost function and first derivative computation. This allows to obtain computationally efficient solution methods. The total least squares family of methods has a wide range of applications in system theory, signal processing, and computer algebra. We describe the applications for deconvolution, linear prediction, and errors-in-variables system identification. Keywords: Total least squares; Orthogonal regression; Errors-in-variables model; Deconvolution; Linear pre- diction; System identification. 1 Introduction The total least squares method was introduced by Golub and Van Loan [25, 27] as a solution technique for an overde- termined system of equations AX B, where A Rm n and B Rm d are the given data and X Rn d is unknown. -
An Analysis of Random Design Linear Regression
An Analysis of Random Design Linear Regression Daniel Hsu1,2, Sham M. Kakade2, and Tong Zhang1 1Department of Statistics, Rutgers University 2Department of Statistics, Wharton School, University of Pennsylvania Abstract The random design setting for linear regression concerns estimators based on a random sam- ple of covariate/response pairs. This work gives explicit bounds on the prediction error for the ordinary least squares estimator and the ridge regression estimator under mild assumptions on the covariate/response distributions. In particular, this work provides sharp results on the \out-of-sample" prediction error, as opposed to the \in-sample" (fixed design) error. Our anal- ysis also explicitly reveals the effect of noise vs. modeling errors. The approach reveals a close connection to the more traditional fixed design setting, and our methods make use of recent ad- vances in concentration inequalities (for vectors and matrices). We also describe an application of our results to fast least squares computations. 1 Introduction In the random design setting for linear regression, one is given pairs (X1;Y1);:::; (Xn;Yn) of co- variates and responses, sampled from a population, where each Xi are random vectors and Yi 2 R. These pairs are hypothesized to have the linear relationship > Yi = Xi β + i for some linear map β, where the i are noise terms. The goal of estimation in this setting is to find coefficients β^ based on these (Xi;Yi) pairs such that the expected prediction error on a new > 2 draw (X; Y ) from the population, measured as E[(X β^ − Y ) ], is as small as possible. -
ROBUST LINEAR LEAST SQUARES REGRESSION 3 Bias Term R(F ∗) R(F (Reg)) Has the Order D/N of the Estimation Term (See [3, 6, 10] and References− Within)
The Annals of Statistics 2011, Vol. 39, No. 5, 2766–2794 DOI: 10.1214/11-AOS918 c Institute of Mathematical Statistics, 2011 ROBUST LINEAR LEAST SQUARES REGRESSION By Jean-Yves Audibert and Olivier Catoni Universit´eParis-Est and CNRS/Ecole´ Normale Sup´erieure/INRIA and CNRS/Ecole´ Normale Sup´erieure and INRIA We consider the problem of robustly predicting as well as the best linear combination of d given functions in least squares regres- sion, and variants of this problem including constraints on the pa- rameters of the linear combination. For the ridge estimator and the ordinary least squares estimator, and their variants, we provide new risk bounds of order d/n without logarithmic factor unlike some stan- dard results, where n is the size of the training data. We also provide a new estimator with better deviations in the presence of heavy-tailed noise. It is based on truncating differences of losses in a min–max framework and satisfies a d/n risk bound both in expectation and in deviations. The key common surprising factor of these results is the absence of exponential moment condition on the output distribution while achieving exponential deviations. All risk bounds are obtained through a PAC-Bayesian analysis on truncated differences of losses. Experimental results strongly back up our truncated min–max esti- mator. 1. Introduction. Our statistical task. Let Z1 = (X1,Y1),...,Zn = (Xn,Yn) be n 2 pairs of input–output and assume that each pair has been independently≥ drawn from the same unknown distribution P . Let denote the input space and let the output space be the set of real numbersX R, so that P is a proba- bility distribution on the product space , R. -
Testing for Heteroskedastic Mixture of Ordinary Least 5
TESTING FOR HETEROSKEDASTIC MIXTURE OF ORDINARY LEAST 5. SQUARES ERRORS Chamil W SENARATHNE1 Wei JIANGUO2 Abstract There is no procedure available in the existing literature to test for heteroskedastic mixture of distributions of residuals drawn from ordinary least squares regressions. This is the first paper that designs a simple test procedure for detecting heteroskedastic mixture of ordinary least squares residuals. The assumption that residuals must be drawn from a homoscedastic mixture of distributions is tested in addition to detecting heteroskedasticity. The test procedure has been designed to account for mixture of distributions properties of the regression residuals when the regressor is drawn with reference to an active market. To retain efficiency of the test, an unbiased maximum likelihood estimator for the true (population) variance was drawn from a log-normal normal family. The results show that there are significant disagreements between the heteroskedasticity detection results of the two auxiliary regression models due to the effect of heteroskedastic mixture of residual distributions. Forecasting exercise shows that there is a significant difference between the two auxiliary regression models in market level regressions than non-market level regressions that supports the new model proposed. Monte Carlo simulation results show significant improvements in the model performance for finite samples with less size distortion. The findings of this study encourage future scholars explore possibility of testing heteroskedastic mixture effect of residuals drawn from multiple regressions and test heteroskedastic mixture in other developed and emerging markets under different market conditions (e.g. crisis) to see the generalisatbility of the model. It also encourages developing other types of tests such as F-test that also suits data generating process. -
Uncorrected Proof
Annals of Operations Research 124, 69–79, 2003 2003 Kluwer Academic Publishers. Manufactured in The Netherlands. 1 1 2 2 3 3 4 Multiple Neutral Data Fitting 4 5 5 6 6 CHRIS TOFALLIS [email protected] 7 7 Deptament of Stats, Econ, Accounting and Management Systems, Business School, 8 University of Hertfordshire, UK 8 9 9 10 10 11 Abstract. A method is proposed for estimating the relationship between a number of variables; this differs 11 12 from regression where the emphasis is on predicting one of the variables. Regression assumes that only 12 13 one of the variables has error or natural variability, whereas our technique does not make this assumption; 13 instead, it treats all variables in the same way and produces models which are units invariant – this is 14 important for ensuring physically meaningful relationships. It is thus superior to orthogonal regression in 14 15 that it does not suffer from being scale-dependent. We show that the solution to the estimation problem is 15 16 a unique and global optimum. For two variables the method has appeared under different names in various 16 17 disciplines, with two Nobel laureates having published work on it. 17 18 18 Keywords: functional relation, data fitting, regression, model fitting 19 19 20 20 21 Introduction 21 22 22 23 23 24 The method of regression is undoubtedly one of the most widely used quantitative meth- 24 25 ods in both the natural and social sciences. For most users of the technique the term is 25 26 taken to refer to the fitting of a function to data by means of the least squares criterion. -
Bayesian Uncertainty Analysis for a Regression Model Versus Application of GUM Supplement 1 to the Least-Squares Estimate
Home Search Collections Journals About Contact us My IOPscience Bayesian uncertainty analysis for a regression model versus application of GUM Supplement 1 to the least-squares estimate This content has been downloaded from IOPscience. Please scroll down to see the full text. 2011 Metrologia 48 233 (http://iopscience.iop.org/0026-1394/48/5/001) View the table of contents for this issue, or go to the journal homepage for more Download details: IP Address: 129.6.222.157 This content was downloaded on 20/02/2014 at 14:34 Please note that terms and conditions apply. IOP PUBLISHING METROLOGIA Metrologia 48 (2011) 233–240 doi:10.1088/0026-1394/48/5/001 Bayesian uncertainty analysis for a regression model versus application of GUM Supplement 1 to the least-squares estimate Clemens Elster1 and Blaza Toman2 1 Physikalisch-Technische Bundesanstalt, Abbestrasse 2-12, 10587 Berlin, Germany 2 Statistical Engineering Division, Information Technology Laboratory, National Institute of Standards and Technology, US Department of Commerce, Gaithersburg, MD, USA Received 7 March 2011, in final form 5 May 2011 Published 16 June 2011 Online at stacks.iop.org/Met/48/233 Abstract Application of least-squares as, for instance, in curve fitting is an important tool of data analysis in metrology. It is tempting to employ the supplement 1 to the GUM (GUM-S1) to evaluate the uncertainty associated with the resulting parameter estimates, although doing so is beyond the specified scope of GUM-S1. We compare the result of such a procedure with a Bayesian uncertainty analysis of the corresponding regression model. -
ROBUST SOLUTIONS to LEAST-SQUARES PROBLEMS with UNCERTAIN DATA ∗ ´ LAURENT EL GHAOUI† and HERVE LEBRET† Abstract
SIAM J. MATRIX ANAL. APPL. c 1997 Society for Industrial and Applied Mathematics Vol. 18, No. 4, pp. 1035–1064, October 1997 015 ROBUST SOLUTIONS TO LEAST-SQUARES PROBLEMS WITH UNCERTAIN DATA ∗ ´ LAURENT EL GHAOUI† AND HERVE LEBRET† Abstract. We consider least-squares problems where the coefficient matrices A, b are unknown but bounded. We minimize the worst-case residual error using (convex) second-order cone program- ming, yielding an algorithm with complexity similar to one singular value decomposition of A. The method can be interpreted as a Tikhonov regularization procedure, with the advantage that it pro- vides an exact bound on the robustness of solution and a rigorous way to compute the regularization parameter. When the perturbation has a known (e.g., Toeplitz) structure, the same problem can be solved in polynomial-time using semidefinite programming (SDP). We also consider the case when A, b are rational functions of an unknown-but-bounded perturbation vector. We show how to mini- mize (via SDP) upper bounds on the optimal worst-case residual. We provide numerical examples, including one from robust identification and one from robust interpolation. Key words. least-squares problems, uncertainty, robustness, second-order cone programming, semidefinite programming, ill-conditioned problem, regularization, robust identification, robust in- terpolation AMS subject classifications. 15A06, 65F10, 65F35, 65K10, 65Y20 PII. S0895479896298130 Notation. For a matrix X, X denotes the largest singular value and X F k k k k the Frobenius norm. If x is a vector, maxi xi is denoted by x . For a matrix A, | | k k∞ A† denotes the Moore–Penrose pseudoinverse of A. -
Chapter 1 Linear Regression with One Predictor Variable
54 Part I Simple Linear Regression 55 Chapter 1 Linear Regression With One Predictor Variable We look at linear regression models with one predictor variable. 1.1 Relations Between Variables Scatter plots of data are often modelled (approximated, simpli¯ed) by linear or non- linear functions. One \good" way to model the data is by using statistical relations. Whereas functions pass through every point in the scatter plot, statistical relations do not necessarily pass through every point but do follow the systematic pattern of the data. Exercise 1.1 (Functional Relation and Statistical Relation) 1. Linear Function Model: Reading Ability Versus Level of Illumination. Consider the following three linear functions used to model the reading ability versus level of illumination data. y = (14/6)x + 69 y = 1.88x + 74.1 y = 6x + 58 B(7,100) B(7,100) 100 100 B(7,100) 100 y y C(9.90) , C(9.90) , y C(9.90) y y , y ilit ilit b b ilit a 80 a 80 b a 80 g g A(3,76) A(3,76) g A(3,76) in in d d in d ea ea r r ea r 60 60 60 0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10 level of illumination, x level of illumination, x level of illumination, x (b) two point BC linear function model (c) least squares linear function model (a) two point AB linear function model Figure 1.1 (Linear Function Models of Reading Ability Versus Level of Illumination) 57 58 Chapter 1. -
Chapter 7 Least Squares Estimation
7-1 Least Squares Estimation Version 1.3 Chapter 7 Least Squares Estimation 7.1. Introduction Least squares is a time-honored estimation procedure, that was developed independently by Gauss (1795), Legendre (1805) and Adrain (1808) and published in the first decade of the nineteenth century. It is perhaps the most widely used technique in geophysical data analysis. Unlike maximum likelihood, which can be applied to any problem for which we know the general form of the joint pdf, in least squares the parameters to be estimated must arise in expressions for the means of the observations. When the parameters appear linearly in these expressions then the least squares estimation problem can be solved in closed form, and it is relatively straightforward to derive the statistical properties for the resulting parameter estimates. One very simple example which we will treat in some detail in order to illustrate the more general problem is that of fitting a straight line to a collection of pairs of observations (xi, yi) where i = 1, 2, . , n. We suppose that a reasonable model is of the form y = β0 + β1x, (1) and we need a mechanism for determining β0 and β1. This is of course just a special case of many more general problems including fitting a polynomial of order p, for which one would need to find p + 1 coefficients. The most commonly used method for finding a model is that of least squares estimation. It is supposed that x is an independent (or predictor) variable which is known exactly, while y is a dependent (or response) variable.