Statistics 3858 : Statistical Models, Parameter Space and Identifiability
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
A Recursive Formula for Moments of a Binomial Distribution Arp´ Ad´ Benyi´ ([email protected]), University of Massachusetts, Amherst, MA 01003 and Saverio M
A Recursive Formula for Moments of a Binomial Distribution Arp´ ad´ Benyi´ ([email protected]), University of Massachusetts, Amherst, MA 01003 and Saverio M. Manago ([email protected]) Naval Postgraduate School, Monterey, CA 93943 While teaching a course in probability and statistics, one of the authors came across an apparently simple question about the computation of higher order moments of a ran- dom variable. The topic of moments of higher order is rarely emphasized when teach- ing a statistics course. The textbooks we came across in our classes, for example, treat this subject rather scarcely; see [3, pp. 265–267], [4, pp. 184–187], also [2, p. 206]. Most of the examples given in these books stop at the second moment, which of course suffices if one is only interested in finding, say, the dispersion (or variance) of a ran- 2 2 dom variable X, D (X) = M2(X) − M(X) . Nevertheless, moments of order higher than 2 are relevant in many classical statistical tests when one assumes conditions of normality. These assumptions may be checked by examining the skewness or kurto- sis of a probability distribution function. The skewness, or the first shape parameter, corresponds to the the third moment about the mean. It describes the symmetry of the tails of a probability distribution. The kurtosis, also known as the second shape pa- rameter, corresponds to the fourth moment about the mean and measures the relative peakedness or flatness of a distribution. Significant skewness or kurtosis indicates that the data is not normal. However, we arrived at higher order moments unintentionally. -
Higher-Order Asymptotics
Higher-Order Asymptotics Todd Kuffner Washington University in St. Louis WHOA-PSI 2016 1 / 113 First- and Higher-Order Asymptotics Classical Asymptotics in Statistics: available sample size n ! 1 First-Order Asymptotic Theory: asymptotic statements that are correct to order O(n−1=2) Higher-Order Asymptotics: refinements to first-order results 1st order 2nd order 3rd order kth order error O(n−1=2) O(n−1) O(n−3=2) O(n−k=2) or or or or o(1) o(n−1=2) o(n−1) o(n−(k−1)=2) Why would anyone care? deeper understanding more accurate inference compare different approaches (which agree to first order) 2 / 113 Points of Emphasis Convergence pointwise or uniform? Error absolute or relative? Deviation region moderate or large? 3 / 113 Common Goals Refinements for better small-sample performance Example Edgeworth expansion (absolute error) Example Barndorff-Nielsen’s R∗ Accurate Approximation Example saddlepoint methods (relative error) Example Laplace approximation Comparative Asymptotics Example probability matching priors Example conditional vs. unconditional frequentist inference Example comparing analytic and bootstrap procedures Deeper Understanding Example sources of inaccuracy in first-order theory Example nuisance parameter effects 4 / 113 Is this relevant for high-dimensional statistical models? The Classical asymptotic regime is when the parameter dimension p is fixed and the available sample size n ! 1. What if p < n or p is close to n? 1. Find a meaningful non-asymptotic analysis of the statistical procedure which works for any n or p (concentration inequalities) 2. Allow both n ! 1 and p ! 1. 5 / 113 Some First-Order Theory Univariate (classical) CLT: Assume X1;X2;::: are i.i.d. -
Stable Components in the Parameter Plane of Transcendental Functions of Finite Type
Stable components in the parameter plane of transcendental functions of finite type N´uriaFagella∗ Dept. de Matem`atiquesi Inform`atica Univ. de Barcelona, Gran Via 585 Barcelona Graduate School of Mathematics (BGSMath) 08007 Barcelona, Spain [email protected] Linda Keen† CUNY Graduate Center 365 Fifth Avenue, New York NY, NY 10016 [email protected], [email protected] November 11, 2020 Abstract We study the parameter planes of certain one-dimensional, dynamically-defined slices of holomorphic families of entire and meromorphic transcendental maps of finite type. Our planes are defined by constraining the orbits of all but one of the singular values, and leaving free one asymptotic value. We study the structure of the regions of parameters, which we call shell components, for which the free asymptotic value tends to an attracting cycle of non-constant multiplier. The exponential and the tangent families are examples that have been studied in detail, and the hyperbolic components in those parameter planes are shell components. Our results apply to slices of both entire and meromorphic maps. We prove that shell components are simply connected, have a locally connected boundary and have no center, i.e., no parameter value for which the cycle is superattracting. Instead, there is a unique parameter in the boundary, the virtual center, which plays the same role. For entire slices, the virtual center is always at infinity, while for meromorphic ones it maybe finite or infinite. In the dynamical plane we prove, among other results, that the basins of attraction which contain only one asymptotic value and no critical points are simply connected. -
Section 7 Testing Hypotheses About Parameters of Normal Distribution. T-Tests and F-Tests
Section 7 Testing hypotheses about parameters of normal distribution. T-tests and F-tests. We will postpone a more systematic approach to hypotheses testing until the following lectures and in this lecture we will describe in an ad hoc way T-tests and F-tests about the parameters of normal distribution, since they are based on a very similar ideas to confidence intervals for parameters of normal distribution - the topic we have just covered. Suppose that we are given an i.i.d. sample from normal distribution N(µ, ν2) with some unknown parameters µ and ν2 : We will need to decide between two hypotheses about these unknown parameters - null hypothesis H0 and alternative hypothesis H1: Hypotheses H0 and H1 will be one of the following: H : µ = µ ; H : µ = µ ; 0 0 1 6 0 H : µ µ ; H : µ < µ ; 0 ∼ 0 1 0 H : µ µ ; H : µ > µ ; 0 ≈ 0 1 0 where µ0 is a given ’hypothesized’ parameter. We will also consider similar hypotheses about parameter ν2 : We want to construct a decision rule α : n H ; H X ! f 0 1g n that given an i.i.d. sample (X1; : : : ; Xn) either accepts H0 or rejects H0 (accepts H1). Null hypothesis is usually a ’main’ hypothesis2 X in a sense that it is expected or presumed to be true and we need a lot of evidence to the contrary to reject it. To quantify this, we pick a parameter � [0; 1]; called level of significance, and make sure that a decision rule α rejects H when it is2 actually true with probability �; i.e. -
Backpropagation TA: Yi Wen
Backpropagation TA: Yi Wen April 17, 2020 CS231n Discussion Section Slides credits: Barak Oshri, Vincent Chen, Nish Khandwala, Yi Wen Agenda ● Motivation ● Backprop Tips & Tricks ● Matrix calculus primer Agenda ● Motivation ● Backprop Tips & Tricks ● Matrix calculus primer Motivation Recall: Optimization objective is minimize loss Motivation Recall: Optimization objective is minimize loss Goal: how should we tweak the parameters to decrease the loss? Agenda ● Motivation ● Backprop Tips & Tricks ● Matrix calculus primer A Simple Example Loss Goal: Tweak the parameters to minimize loss => minimize a multivariable function in parameter space A Simple Example => minimize a multivariable function Plotted on WolframAlpha Approach #1: Random Search Intuition: the step we take in the domain of function Approach #2: Numerical Gradient Intuition: rate of change of a function with respect to a variable surrounding a small region Approach #2: Numerical Gradient Intuition: rate of change of a function with respect to a variable surrounding a small region Finite Differences: Approach #3: Analytical Gradient Recall: partial derivative by limit definition Approach #3: Analytical Gradient Recall: chain rule Approach #3: Analytical Gradient Recall: chain rule E.g. Approach #3: Analytical Gradient Recall: chain rule E.g. Approach #3: Analytical Gradient Recall: chain rule Intuition: upstream gradient values propagate backwards -- we can reuse them! Gradient “direction and rate of fastest increase” Numerical Gradient vs Analytical Gradient What about Autograd? -
Decomposing the Parameter Space of Biological Networks Via a Numerical Discriminant Approach
Decomposing the parameter space of biological networks via a numerical discriminant approach Heather A. Harrington1, Dhagash Mehta2, Helen M. Byrne1, and Jonathan D. Hauenstein2 1 Mathematical Institute, The University of Oxford, Oxford OX2 6GG, UK {harrington,helen.byrne}@maths.ox.ac.uk, www.maths.ox.ac.uk/people/{heather.harrington,helen.byrne} 2 Department of Applied and Computational Mathematics and Statistics, University of Notre Dame, Notre Dame IN 46556, USA {dmehta,hauenstein}@nd.edu, www.nd.edu/~{dmehta,jhauenst} Abstract. Many systems in biology (as well as other physical and en- gineering systems) can be described by systems of ordinary dierential equation containing large numbers of parameters. When studying the dynamic behavior of these large, nonlinear systems, it is useful to iden- tify and characterize the steady-state solutions as the model parameters vary, a technically challenging problem in a high-dimensional parameter landscape. Rather than simply determining the number and stability of steady-states at distinct points in parameter space, we decompose the parameter space into nitely many regions, the number and structure of the steady-state solutions being consistent within each distinct region. From a computational algebraic viewpoint, the boundary of these re- gions is contained in the discriminant locus. We develop global and local numerical algorithms for constructing the discriminant locus and classi- fying the parameter landscape. We showcase our numerical approaches by applying them to molecular and cell-network models. Keywords: parameter landscape · numerical algebraic geometry · dis- criminant locus · cellular networks. 1 Introduction The dynamic behavior of many biophysical systems can be mathematically mod- eled with systems of dierential equations that describe how the state variables interact and evolve over time. -
The Method of Maximum Likelihood for Simple Linear Regression
08:48 Saturday 19th September, 2015 See updates and corrections at http://www.stat.cmu.edu/~cshalizi/mreg/ Lecture 6: The Method of Maximum Likelihood for Simple Linear Regression 36-401, Fall 2015, Section B 17 September 2015 1 Recapitulation We introduced the method of maximum likelihood for simple linear regression in the notes for two lectures ago. Let's review. We start with the statistical model, which is the Gaussian-noise simple linear regression model, defined as follows: 1. The distribution of X is arbitrary (and perhaps X is even non-random). 2. If X = x, then Y = β0 + β1x + , for some constants (\coefficients", \parameters") β0 and β1, and some random noise variable . 3. ∼ N(0; σ2), and is independent of X. 4. is independent across observations. A consequence of these assumptions is that the response variable Y is indepen- dent across observations, conditional on the predictor X, i.e., Y1 and Y2 are independent given X1 and X2 (Exercise 1). As you'll recall, this is a special case of the simple linear regression model: the first two assumptions are the same, but we are now assuming much more about the noise variable : it's not just mean zero with constant variance, but it has a particular distribution (Gaussian), and everything we said was uncorrelated before we now strengthen to independence1. Because of these stronger assumptions, the model tells us the conditional pdf 2 of Y for each x, p(yjX = x; β0; β1; σ ). (This notation separates the random variables from the parameters.) Given any data set (x1; y1); (x2; y2);::: (xn; yn), we can now write down the probability density, under the model, of seeing that data: n n (y −(β +β x ))2 Y 2 Y 1 − i 0 1 i p(yijxi; β0; β1; σ ) = p e 2σ2 2 i=1 i=1 2πσ 1See the notes for lecture 1 for a reminder, with an explicit example, of how uncorrelated random variables can nonetheless be strongly statistically dependent. -
The Emergence of Gravitational Wave Science: 100 Years of Development of Mathematical Theory, Detectors, Numerical Algorithms, and Data Analysis Tools
BULLETIN (New Series) OF THE AMERICAN MATHEMATICAL SOCIETY Volume 53, Number 4, October 2016, Pages 513–554 http://dx.doi.org/10.1090/bull/1544 Article electronically published on August 2, 2016 THE EMERGENCE OF GRAVITATIONAL WAVE SCIENCE: 100 YEARS OF DEVELOPMENT OF MATHEMATICAL THEORY, DETECTORS, NUMERICAL ALGORITHMS, AND DATA ANALYSIS TOOLS MICHAEL HOLST, OLIVIER SARBACH, MANUEL TIGLIO, AND MICHELE VALLISNERI In memory of Sergio Dain Abstract. On September 14, 2015, the newly upgraded Laser Interferometer Gravitational-wave Observatory (LIGO) recorded a loud gravitational-wave (GW) signal, emitted a billion light-years away by a coalescing binary of two stellar-mass black holes. The detection was announced in February 2016, in time for the hundredth anniversary of Einstein’s prediction of GWs within the theory of general relativity (GR). The signal represents the first direct detec- tion of GWs, the first observation of a black-hole binary, and the first test of GR in its strong-field, high-velocity, nonlinear regime. In the remainder of its first observing run, LIGO observed two more signals from black-hole bina- ries, one moderately loud, another at the boundary of statistical significance. The detections mark the end of a decades-long quest and the beginning of GW astronomy: finally, we are able to probe the unseen, electromagnetically dark Universe by listening to it. In this article, we present a short historical overview of GW science: this young discipline combines GR, arguably the crowning achievement of classical physics, with record-setting, ultra-low-noise laser interferometry, and with some of the most powerful developments in the theory of differential geometry, partial differential equations, high-performance computation, numerical analysis, signal processing, statistical inference, and data science. -
Use of the Kurtosis Statistic in the Frequency Domain As an Aid In
lEEE JOURNALlEEE OF OCEANICENGINEERING, VOL. OE-9, NO. 2, APRIL 1984 85 Use of the Kurtosis Statistic in the FrequencyDomain as an Aid in Detecting Random Signals Absmact-Power spectral density estimation is often employed as a couldbe utilized in signal processing. The objective ofthis method for signal ,detection. For signals which occur randomly, a paper is to compare the PSD technique for signal processing frequency domain kurtosis estimate supplements the power spectral witha new methodwhich computes the frequency domain density estimate and, in some cases, can be.employed to detect their presence. This has been verified from experiments vith real data of kurtosis (FDK) [2] forthe real and imaginary parts of the randomly occurring signals. In order to better understand the detec- complex frequency components. Kurtosis is defined as a ratio tion of randomlyoccurring signals, sinusoidal and narrow-band of a fourth-order central moment to the square of a second- Gaussian signals are considered, which when modeled to represent a order central moment. fading or multipath environment, are received as nowGaussian in Using theNeyman-Pearson theory in thetime domain, terms of a frequency domain kurtosis estimate. Several fading and multipath propagation probability density distributions of practical Ferguson [3] , has shown that kurtosis is a locally optimum interestare considered, including Rayleigh and log-normal. The detectionstatistic under certain conditions. The reader is model is generalized to handle transient and frequency modulated referred to Ferguson'swork for the details; however, it can signals by taking into account the probability of the signal being in a be simply said thatit is concernedwith detecting outliers specific frequency range over the total data interval. -
11. Parameter Estimation
11. Parameter Estimation Chris Piech and Mehran Sahami May 2017 We have learned many different distributions for random variables and all of those distributions had parame- ters: the numbers that you provide as input when you define a random variable. So far when we were working with random variables, we either were explicitly told the values of the parameters, or, we could divine the values by understanding the process that was generating the random variables. What if we don’t know the values of the parameters and we can’t estimate them from our own expert knowl- edge? What if instead of knowing the random variables, we have a lot of examples of data generated with the same underlying distribution? In this chapter we are going to learn formal ways of estimating parameters from data. These ideas are critical for artificial intelligence. Almost all modern machine learning algorithms work like this: (1) specify a probabilistic model that has parameters. (2) Learn the value of those parameters from data. Parameters Before we dive into parameter estimation, first let’s revisit the concept of parameters. Given a model, the parameters are the numbers that yield the actual distribution. In the case of a Bernoulli random variable, the single parameter was the value p. In the case of a Uniform random variable, the parameters are the a and b values that define the min and max value. Here is a list of random variables and the corresponding parameters. From now on, we are going to use the notation q to be a vector of all the parameters: Distribution Parameters Bernoulli(p) q = p Poisson(l) q = l Uniform(a,b) q = (a;b) Normal(m;s 2) q = (m;s 2) Y = mX + b q = (m;b) In the real world often you don’t know the “true” parameters, but you get to observe data. -
A Widely Applicable Bayesian Information Criterion
JournalofMachineLearningResearch14(2013)867-897 Submitted 8/12; Revised 2/13; Published 3/13 A Widely Applicable Bayesian Information Criterion Sumio Watanabe [email protected] Department of Computational Intelligence and Systems Science Tokyo Institute of Technology Mailbox G5-19, 4259 Nagatsuta, Midori-ku Yokohama, Japan 226-8502 Editor: Manfred Opper Abstract A statistical model or a learning machine is called regular if the map taking a parameter to a prob- ability distribution is one-to-one and if its Fisher information matrix is always positive definite. If otherwise, it is called singular. In regular statistical models, the Bayes free energy, which is defined by the minus logarithm of Bayes marginal likelihood, can be asymptotically approximated by the Schwarz Bayes information criterion (BIC), whereas in singular models such approximation does not hold. Recently, it was proved that the Bayes free energy of a singular model is asymptotically given by a generalized formula using a birational invariant, the real log canonical threshold (RLCT), instead of half the number of parameters in BIC. Theoretical values of RLCTs in several statistical models are now being discovered based on algebraic geometrical methodology. However, it has been difficult to estimate the Bayes free energy using only training samples, because an RLCT depends on an unknown true distribution. In the present paper, we define a widely applicable Bayesian information criterion (WBIC) by the average log likelihood function over the posterior distribution with the inverse temperature 1/logn, where n is the number of training samples. We mathematically prove that WBIC has the same asymptotic expansion as the Bayes free energy, even if a statistical model is singular for or unrealizable by a statistical model. -
Statistic: a Quantity That We Can Calculate from Sample Data That Summarizes a Characteristic of That Sample
STAT 509 – Section 4.1 – Estimation Parameter: A numerical characteristic of a population. Examples: Statistic: A quantity that we can calculate from sample data that summarizes a characteristic of that sample. Examples: Point Estimator: A statistic which is a single number meant to estimate a parameter. It would be nice if the average value of the estimator (over repeated sampling) equaled the target parameter. An estimator is called unbiased if the mean of its sampling distribution is equal to the parameter being estimated. Examples: Another nice property of an estimator: we want it to be as precise as possible. The standard deviation of a statistic’s sampling distribution is called the standard error of the statistic. The standard error of the sample mean Y is / n . Note: As the sample size gets larger, the spread of the sampling distribution gets smaller. When the sample size is large, the sample mean varies less across samples. Evaluating an estimator: (1) Is it unbiased? (2) Does it have a small standard error? Interval Estimates • With a point estimate, we used a single number to estimate a parameter. • We can also use a set of numbers to serve as “reasonable” estimates for the parameter. Example: Assume we have a sample of size n from a normally distributed population. Y T We know: s / n has a t-distribution with n – 1 degrees of freedom. (Exactly true when data are normal, approximately true when data non-normal but n is large.) Y P(t t ) So: 1 – = n1, / 2 s / n n1, / 2 = where tn–1, /2 = the t-value with /2 area to the right (can be found from Table 2) This formula is called a “confidence interval” for .