STAT 830 the Basics of Nonparametric Models The
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Von Mises' Frequentist Approach to Probability
Section on Statistical Education – JSM 2008 Von Mises’ Frequentist Approach to Probability Milo Schield1 and Thomas V. V. Burnham2 1 Augsburg College, Minneapolis, MN 2 Cognitive Systems, San Antonio, TX Abstract Richard Von Mises (1883-1953), the originator of the birthday problem, viewed statistics and probability theory as a science that deals with and is based on real objects rather than as a branch of mathematics that simply postulates relationships from which conclusions are derived. In this context, he formulated a strict Frequentist approach to probability. This approach is limited to observations where there are sufficient reasons to project future stability – to believe that the relative frequency of the observed attributes would tend to a fixed limit if the observations were continued indefinitely by sampling from a collective. This approach is reviewed. Some suggestions are made for statistical education. 1. Richard von Mises Richard von Mises (1883-1953) may not be well-known by statistical educators even though in 1939 he first proposed a problem that is today widely-known as the birthday problem.1 There are several reasons for this lack of recognition. First, he was educated in engineering and worked in that area for his first 30 years. Second, he published primarily in German. Third, his works in statistics focused more on theoretical foundations. Finally, his inductive approach to deriving the basic operations of probability is very different from the postulate-and-derive approach normally taught in introductory statistics. See Frank (1954) and Studies in Mathematics and Mechanics (1954) for a review of von Mises’ works. This paper is based on his two English-language books on probability: Probability, Statistics and Truth (1936) and Mathematical Theory of Probability and Statistics (1964). -
Phase Transition Unbiased Estimation in High Dimensional Settings Arxiv
Phase Transition Unbiased Estimation in High Dimensional Settings St´ephaneGuerrier, Mucyo Karemera, Samuel Orso & Maria-Pia Victoria-Feser Research Center for Statistics, University of Geneva Abstract An important challenge in statistical analysis concerns the control of the finite sample bias of estimators. This problem is magnified in high dimensional settings where the number of variables p diverge with the sample size n. However, it is difficult to establish whether an estimator θ^ of θ0 is unbiased and the asymptotic ^ order of E[θ] − θ0 is commonly used instead. We introduce a new property to assess the bias, called phase transition unbiasedness, which is weaker than unbiasedness but stronger than asymptotic results. An estimator satisfying this property is such ^ ∗ that E[θ] − θ0 2 = 0, for all n greater than a finite sample size n . We propose a phase transition unbiased estimator by matching an initial estimator computed on the sample and on simulated data. It is computed using an algorithm which is shown to converge exponentially fast. The initial estimator is not required to be consistent and thus may be conveniently chosen for computational efficiency or for other properties. We demonstrate the consistency and the limiting distribution of the estimator in high dimension. Finally, we develop new estimators for logistic regression models, with and without random effects, that enjoy additional properties such as robustness to data contamination and to the problem of separability. arXiv:1907.11541v3 [math.ST] 1 Nov 2019 Keywords: Finite sample bias, Iterative bootstrap, Two-step estimators, Indirect inference, Robust estimation, Logistic regression. 1 1. Introduction An important challenge in statistical analysis concerns the control of the (finite sample) bias of estimators. -
Autocorrelation and Frequency-Resolved Optical Gating Measurements Based on the Third Harmonic Generation in a Gaseous Medium
Appl. Sci. 2015, 5, 136-144; doi:10.3390/app5020136 OPEN ACCESS applied sciences ISSN 2076-3417 www.mdpi.com/journal/applsci Article Autocorrelation and Frequency-Resolved Optical Gating Measurements Based on the Third Harmonic Generation in a Gaseous Medium Yoshinari Takao 1,†, Tomoko Imasaka 2,†, Yuichiro Kida 1,† and Totaro Imasaka 1,3,* 1 Department of Applied Chemistry, Graduate School of Engineering, Kyushu University, 744, Motooka, Nishi-ku, Fukuoka 819-0395, Japan; E-Mails: [email protected] (Y.T.); [email protected] (Y.K.) 2 Laboratory of Chemistry, Graduate School of Design, Kyushu University, 4-9-1, Shiobaru, Minami-ku, Fukuoka 815-8540, Japan; E-Mail: [email protected] 3 Division of Optoelectronics and Photonics, Center for Future Chemistry, Kyushu University, 744, Motooka, Nishi-ku, Fukuoka 819-0395, Japan † These authors contributed equally to this work. * Author to whom correspondence should be addressed; E-Mail: [email protected]; Tel.: +81-92-802-2883; Fax: +81-92-802-2888. Academic Editor: Takayoshi Kobayashi Received: 7 April 2015 / Accepted: 3 June 2015 / Published: 9 June 2015 Abstract: A gas was utilized in producing the third harmonic emission as a nonlinear optical medium for autocorrelation and frequency-resolved optical gating measurements to evaluate the pulse width and chirp of a Ti:sapphire laser. Due to a wide frequency domain available for a gas, this approach has potential for use in measuring the pulse width in the optical (ultraviolet/visible) region beyond one octave and thus for measuring an optical pulse width less than 1 fs. -
FORMULAS from EPIDEMIOLOGY KEPT SIMPLE (3E) Chapter 3: Epidemiologic Measures
FORMULAS FROM EPIDEMIOLOGY KEPT SIMPLE (3e) Chapter 3: Epidemiologic Measures Basic epidemiologic measures used to quantify: • frequency of occurrence • the effect of an exposure • the potential impact of an intervention. Epidemiologic Measures Measures of disease Measures of Measures of potential frequency association impact (“Measures of Effect”) Incidence Prevalence Absolute measures of Relative measures of Attributable Fraction Attributable Fraction effect effect in exposed cases in the Population Incidence proportion Incidence rate Risk difference Risk Ratio (Cumulative (incidence density, (Incidence proportion (Incidence Incidence, Incidence hazard rate, person- difference) Proportion Ratio) Risk) time rate) Incidence odds Rate Difference Rate Ratio (Incidence density (Incidence density difference) ratio) Prevalence Odds Ratio Difference Macintosh HD:Users:buddygerstman:Dropbox:eks:formula_sheet.doc Page 1 of 7 3.1 Measures of Disease Frequency No. of onsets Incidence Proportion = No. at risk at beginning of follow-up • Also called risk, average risk, and cumulative incidence. • Can be measured in cohorts (closed populations) only. • Requires follow-up of individuals. No. of onsets Incidence Rate = ∑person-time • Also called incidence density and average hazard. • When disease is rare (incidence proportion < 5%), incidence rate ≈ incidence proportion. • In cohorts (closed populations), it is best to sum individual person-time longitudinally. It can also be estimated as Σperson-time ≈ (average population size) × (duration of follow-up). Actuarial adjustments may be needed when the disease outcome is not rare. • In an open populations, Σperson-time ≈ (average population size) × (duration of follow-up). Examples of incidence rates in open populations include: births Crude birth rate (per m) = × m mid-year population size deaths Crude mortality rate (per m) = × m mid-year population size deaths < 1 year of age Infant mortality rate (per m) = × m live births No. -
Bias, Mean-Square Error, Relative Efficiency
3 Evaluating the Goodness of an Estimator: Bias, Mean-Square Error, Relative Efficiency Consider a population parameter ✓ for which estimation is desired. For ex- ample, ✓ could be the population mean (traditionally called µ) or the popu- lation variance (traditionally called σ2). Or it might be some other parame- ter of interest such as the population median, population mode, population standard deviation, population minimum, population maximum, population range, population kurtosis, or population skewness. As previously mentioned, we will regard parameters as numerical charac- teristics of the population of interest; as such, a parameter will be a fixed number, albeit unknown. In Stat 252, we will assume that our population has a distribution whose density function depends on the parameter of interest. Most of the examples that we will consider in Stat 252 will involve continuous distributions. Definition 3.1. An estimator ✓ˆ is a statistic (that is, it is a random variable) which after the experiment has been conducted and the data collected will be used to estimate ✓. Since it is true that any statistic can be an estimator, you might ask why we introduce yet another word into our statistical vocabulary. Well, the answer is quite simple, really. When we use the word estimator to describe a particular statistic, we already have a statistical estimation problem in mind. For example, if ✓ is the population mean, then a natural estimator of ✓ is the sample mean. If ✓ is the population variance, then a natural estimator of ✓ is the sample variance. More specifically, suppose that Y1,...,Yn are a random sample from a population whose distribution depends on the parameter ✓.The following estimators occur frequently enough in practice that they have special notations. -
Fundamental Frequency Detection
Fundamental Frequency Detection Jan Cernock´y,ˇ Valentina Hubeika cernocky|ihubeika @fit.vutbr.cz { } DCGM FIT BUT Brno Fundamental Frequency Detection Jan Cernock´y,ValentinaHubeika,DCGMFITBUTBrnoˇ 1/37 Agenda Fundamental frequency characteristics. • Issues. • Autocorrelation method, AMDF, NFFC. • Formant impact reduction. • Long time predictor. • Cepstrum. • Improvements in fundamental frequency detection. • Fundamental Frequency Detection Jan Cernock´y,ValentinaHubeika,DCGMFITBUTBrnoˇ 2/37 Recap – speech production and its model Fundamental Frequency Detection Jan Cernock´y,ValentinaHubeika,DCGMFITBUTBrnoˇ 3/37 Introduction Fundamental frequency, pitch, is the frequency vocal cords oscillate on: F . • 0 The period of fundamental frequency, (pitch period) is T = 1 . • 0 F0 The term “lag” denotes the pitch period expressed in samples : L = T Fs, where Fs • 0 is the sampling frequency. Fundamental Frequency Detection Jan Cernock´y,ValentinaHubeika,DCGMFITBUTBrnoˇ 4/37 Fundamental Frequency Utilization speech synthesis – melody generation. • coding • – in simple encoding such as LPC, reduction of the bit-stream can be reached by separate transfer of the articulatory tract parameters, energy, voiced/unvoiced sound flag and pitch F0. – in more complex encoders (such as RPE-LTP or ACELP in the GSM cell phones) long time predictor LTP is used. LPT is a filter with a “long” impulse response which however contains only few non-zero components. Fundamental Frequency Detection Jan Cernock´y,ValentinaHubeika,DCGMFITBUTBrnoˇ 5/37 Fundamental Frequency Characteristics F takes the values from 50 Hz (males) to 400 Hz (children), with Fs=8000 Hz these • 0 frequencies correspond to the lags L=160 to 20 samples. It can be seen, that with low values F0 approaches the frame length (20 ms, which corresponds to 160 samples). -
ACTS 4304 FORMULA SUMMARY Lesson 1: Basic Probability Summary of Probability Concepts Probability Functions
ACTS 4304 FORMULA SUMMARY Lesson 1: Basic Probability Summary of Probability Concepts Probability Functions F (x) = P r(X ≤ x) S(x) = 1 − F (x) dF (x) f(x) = dx H(x) = − ln S(x) dH(x) f(x) h(x) = = dx S(x) Functions of random variables Z 1 Expected Value E[g(x)] = g(x)f(x)dx −∞ 0 n n-th raw moment µn = E[X ] n n-th central moment µn = E[(X − µ) ] Variance σ2 = E[(X − µ)2] = E[X2] − µ2 µ µ0 − 3µ0 µ + 2µ3 Skewness γ = 3 = 3 2 1 σ3 σ3 µ µ0 − 4µ0 µ + 6µ0 µ2 − 3µ4 Kurtosis γ = 4 = 4 3 2 2 σ4 σ4 Moment generating function M(t) = E[etX ] Probability generating function P (z) = E[zX ] More concepts • Standard deviation (σ) is positive square root of variance • Coefficient of variation is CV = σ/µ • 100p-th percentile π is any point satisfying F (π−) ≤ p and F (π) ≥ p. If F is continuous, it is the unique point satisfying F (π) = p • Median is 50-th percentile; n-th quartile is 25n-th percentile • Mode is x which maximizes f(x) (n) n (n) • MX (0) = E[X ], where M is the n-th derivative (n) PX (0) • n! = P r(X = n) (n) • PX (1) is the n-th factorial moment of X. Bayes' Theorem P r(BjA)P r(A) P r(AjB) = P r(B) fY (yjx)fX (x) fX (xjy) = fY (y) Law of total probability 2 If Bi is a set of exhaustive (in other words, P r([iBi) = 1) and mutually exclusive (in other words P r(Bi \ Bj) = 0 for i 6= j) events, then for any event A, X X P r(A) = P r(A \ Bi) = P r(Bi)P r(AjBi) i i Correspondingly, for continuous distributions, Z P r(A) = P r(Ajx)f(x)dx Conditional Expectation Formula EX [X] = EY [EX [XjY ]] 3 Lesson 2: Parametric Distributions Forms of probability -
Pearson-Fisher Chi-Square Statistic Revisited
Information 2011 , 2, 528-545; doi:10.3390/info2030528 OPEN ACCESS information ISSN 2078-2489 www.mdpi.com/journal/information Communication Pearson-Fisher Chi-Square Statistic Revisited Sorana D. Bolboac ă 1, Lorentz Jäntschi 2,*, Adriana F. Sestra ş 2,3 , Radu E. Sestra ş 2 and Doru C. Pamfil 2 1 “Iuliu Ha ţieganu” University of Medicine and Pharmacy Cluj-Napoca, 6 Louis Pasteur, Cluj-Napoca 400349, Romania; E-Mail: [email protected] 2 University of Agricultural Sciences and Veterinary Medicine Cluj-Napoca, 3-5 M ănăş tur, Cluj-Napoca 400372, Romania; E-Mails: [email protected] (A.F.S.); [email protected] (R.E.S.); [email protected] (D.C.P.) 3 Fruit Research Station, 3-5 Horticultorilor, Cluj-Napoca 400454, Romania * Author to whom correspondence should be addressed; E-Mail: [email protected]; Tel: +4-0264-401-775; Fax: +4-0264-401-768. Received: 22 July 2011; in revised form: 20 August 2011 / Accepted: 8 September 2011 / Published: 15 September 2011 Abstract: The Chi-Square test (χ2 test) is a family of tests based on a series of assumptions and is frequently used in the statistical analysis of experimental data. The aim of our paper was to present solutions to common problems when applying the Chi-square tests for testing goodness-of-fit, homogeneity and independence. The main characteristics of these three tests are presented along with various problems related to their application. The main problems identified in the application of the goodness-of-fit test were as follows: defining the frequency classes, calculating the X2 statistic, and applying the χ2 test. -
Weak Instruments and Finite-Sample Bias
7 Weak instruments and finite-sample bias In this chapter, we consider the effect of weak instruments on instrumen- tal variable (IV) analyses. Weak instruments, which were introduced in Sec- tion 4.5.2, are those that do not explain a large proportion of the variation in the exposure, and so the statistical association between the IV and the expo- sure is not strong. This is of particular relevance in Mendelian randomization studies since the associations of genetic variants with exposures of interest are often weak. This chapter focuses on the impact of weak instruments on the bias and coverage of IV estimates. 7.1 Introduction Although IV techniques can be used to give asymptotically unbiased estimates of causal effects in the presence of confounding, these estimates suffer from bias when evaluated in finite samples [Nelson and Startz, 1990]. A weak instrument (or a weak IV) is still a valid IV, in that it satisfies the IV assumptions, and es- timates using the IV with an infinite sample size will be unbiased; but for any finite sample size, the average value of the IV estimator will be biased. This bias, known as weak instrument bias, is towards the observational confounded estimate. Its magnitude depends on the strength of association between the IV and the exposure, which is measured by the F statistic in the regression of the exposure on the IV [Bound et al., 1995]. In this chapter, we assume the context of ‘one-sample’ Mendelian randomization, in which evidence on the genetic variant, exposure, and outcome are taken on the same set of indi- viduals, rather than subsample (Section 8.5.2) or two-sample (Section 9.8.2) Mendelian randomization, in which genetic associations with the exposure and outcome are estimated in different sets of individuals (overlapping sets in subsample, non-overlapping sets in two-sample Mendelian randomization). -
Approximate Maximum Likelihood Method for Frequency Estimation
Statistica Sinica 10(2000), 157-171 APPROXIMATE MAXIMUM LIKELIHOOD METHOD FOR FREQUENCY ESTIMATION Dawei Huang Queensland University of Technology Abstract: A frequency can be estimated by few Discrete Fourier Transform (DFT) coefficients, see Rife and Vincent (1970), Quinn (1994, 1997). This approach is computationally efficient. However, the statistical efficiency of the estimator de- pends on the location of the frequency. In this paper, we explain this approach from a point of view of an Approximate Maximum Likelihood (AML) method. Then we enhance the efficiency of this method by using more DFT coefficients. Compared to 30% and 61% efficiency in the worst cases in Quinn (1994) and Quinn (1997), respectively, we show that if 13 or 25 DFT coefficients are used, AML will achieve at least 90% or 95% efficiency for all frequency locations. Key words and phrases: Approximate Maximum Likelihood, discrete Fourier trans- form, efficiency, fast algorithm, frequency estimation, semi-sufficient statistics. 1. Introduction A model that has been discussed widely in Time Series Analysis and Signal Processing is the following: xt = Acos(tω + φ)+ut,t=0,...,T − 1, (1) where 0 <A,0 <ω<π,and {ut} is a stationary sequence satisfying certain conditions to be specified later. The objective is to estimate the frequency ω from the observations {xt}. After obtaining an estimator of the frequency, the amplitude A and the phase φ can be estimated by a standard linear least squares method. Two topics have been of interest in the study of this model: estimation efficiency and computational complexity. Following the idea of Best Asymp- totic Normal estimation (BAN), the efficiency hereafter is measured by the ratio of the Cramer - Rao Bound (CRB) for unbiased estimators over the variance of the asymptotic distribution of the estimator. -
Sketching for M-Estimators: a Unified Approach to Robust Regression
Sketching for M-Estimators: A Unified Approach to Robust Regression Kenneth L. Clarkson∗ David P. Woodruffy Abstract ing algorithm often results, since a single pass over A We give algorithms for the M-estimators minx kAx − bkG, suffices to compute SA. n×d n n where A 2 R and b 2 R , and kykG for y 2 R is specified An important property of many of these sketching ≥0 P by a cost function G : R 7! R , with kykG ≡ i G(yi). constructions is that S is a subspace embedding, meaning d The M-estimators generalize `p regression, for which G(x) = that for all x 2 R , kSAxk ≈ kAxk. (Here the vector jxjp. We first show that the Huber measure can be computed norm is generally `p for some p.) For the regression up to relative error in O(nnz(A) log n + poly(d(log n)=")) d time, where nnz(A) denotes the number of non-zero entries problem of minimizing kAx − bk with respect to x 2 R , of the matrix A. Huber is arguably the most widely used for inputs A 2 Rn×d and b 2 Rn, a minor extension of M-estimator, enjoying the robustness properties of `1 as well the embedding condition implies S preserves the norm as the smoothness properties of ` . 2 of the residual vector Ax − b, that is kS(Ax − b)k ≈ We next develop algorithms for general M-estimators. kAx − bk, so that a vector x that makes kS(Ax − b)k We analyze the M-sketch, which is a variation of a sketch small will also make kAx − bk small. -
Bias in Parametric Estimation: Reduction and Useful Side-Effects
Bias in parametric estimation: reduction and useful side-effects Ioannis Kosmidis Department of Statistical Science, University College London August 30, 2018 Abstract The bias of an estimator is defined as the difference of its expected value from the parameter to be estimated, where the expectation is with respect to the model. Loosely speaking, small bias reflects the desire that if an experiment is repeated in- definitely then the average of all the resultant estimates will be close to the parameter value that is estimated. The current paper is a review of the still-expanding repository of methods that have been developed to reduce bias in the estimation of parametric models. The review provides a unifying framework where all those methods are seen as attempts to approximate the solution of a simple estimating equation. Of partic- ular focus is the maximum likelihood estimator, which despite being asymptotically unbiased under the usual regularity conditions, has finite-sample bias that can re- sult in significant loss of performance of standard inferential procedures. An informal comparison of the methods is made revealing some useful practical side-effects in the estimation of popular models in practice including: i) shrinkage of the estimators in binomial and multinomial regression models that guarantees finiteness even in cases of data separation where the maximum likelihood estimator is infinite, and ii) inferential benefits for models that require the estimation of dispersion or precision parameters. Keywords: jackknife/bootstrap, indirect inference, penalized likelihood, infinite es- timates, separation in models with categorical responses 1 Impact of bias in estimation By its definition, bias necessarily depends on how the model is written in terms of its parameters and this dependence makes it not a strong statistical principle in terms of evaluating the performance of estimators; for example, unbiasedness of the familiar sample variance S2 as an estimator of σ2 does not deliver an unbiased estimator of σ itself.