5 Decision Theory: Basic Concepts

Total Page:16

File Type:pdf, Size:1020Kb

5 Decision Theory: Basic Concepts 5 Decision Theory: Basic Concepts Point estimation of an unknown parameter is generally considered the most basic inference problem. Speaking generically, if θ is some unknown parameter taking values in a suitable parameter space Θ, then a point estimate is an educated guess at the true value of θ. Now, of course, we do not just guess the true value of θ without some information. Information comes from data; some information may also come from expert opinion separate from data. In any case, a point estimate is a function of the available sample data. To start with, we allow any function of the data as a possible point estimate. Theory of inference is used to separate the good estimates from the not so good or bad estimates. As usual, we will start with an example to help us understand the general definitions. iid Example 5.1. Suppose X1, ··· , Xn ∼ Poi(λ),λ > 0. Suppose we want to estimate the parameter λ. Now, of course, λ = Eλ(X1), i.e., λ is the population mean. So, just ¯ X1+···+Xn instinctively, we may want to estimate λ by the sample mean X = n ; and, indeed, X¯ is a possible point estimator of λ. While λ takes values in Θ = (0, ∞), the estimator X¯ takes values in A = [0, ∞); X¯ can be equal to zero! So, the set of possible values of the parameter and the set of possible values of an estimator need not be identical. We must allow them to be different sets, in general. Now, X¯ is certainly not the only possible estimator of λ. We can use essentially any function of the sample observations X1, ··· , Xn to estimate the parameter λ. For example, 4 just X1, or X1 + X2 − X3, or even seemingly poor estimators like X1 . Any estimator is allowed to begin with; theory will separate the good ones from the bad ones. iid 2 Next, suppose X1, ··· , Xn ∼ N(µ, σ ), where µ, σ are unknown parameters. So, now, we have got a two dimensional parameter vector, θ = (µ, σ). Suppose we want to estimate µ. Once again, a possible estimator is X¯; a few other possible estimators are the sample X−2+X4 median, Mn = median{X1, ··· , Xn}, or 2 , or even seemingly poor estimators, like 100X¯. Suppose, it was known to us that µ must be nonnegative. Then, the set of possible values of µ is Θ=[0, ∞). However, the instinctive point estimator X¯ can take any real value. It takes values in the set A = (−∞, ∞). You would notice again that A is not the same as Θ in this case. In general, A and Θ can be different sets. If we want instead to estimate σ2, which is the population variance, a first thought is to 2 1 n ¯ 2 use the sample variance s = n−1 i=1(Xi − X) (dividing by n − 1 rather than n seems a little odd at first glance, but hasP a mathematical reason, which will be clear soon). In this example, the parameter σ2 and the estimator s2 both take values in (0, ∞). But, if we knew that σ2 ≤ 100, say, then once again, Θ and A would be different. And, as always, 2 1 n 2 there are many other possible estimators of σ , for example, n−1 i=1(Xi − Mn) , where P Mn is the sample median. Here is a formal definition of a point estimator. 267 (n) Definition 5.1. Let the vector of sample observations X = (X1, ··· , Xn) have a joint p distribution P = Pn and let θ = h(P ), taking values in the parameter space Θ ⊆ R , be a parameter of the distribution P . Let T (X1, ··· , Xn) taking values in a specified set p A ⊆ R be a general statistic, Then, any such T (X1, ··· , Xn) is called a point estimator of θ. The set Θ is called the parameter space, and the set A is called the statistician’s action space. If specific observed sample data X1 = x1, ··· , Xn = xn are available, then the particular value T (x1, ··· ,xn) is called an estimate of θ. Thus, the word estimator applies to the general function T (X1, ··· , Xn), and the word estimate applies to the value T (x1, ··· ,xn) for specific data. In this text, we use estimator and estimate synonymously. A standard general notation for a generic estimate of a parameter θ is θˆ = θˆ(X1, ··· , Xn). 5.0.1 Evaluating an Estimator and MSE Except in rare cases, the estimate θˆ would not be exactly equal to the true value of the unknown parameter θ. It seems reasonable that we like an estimate θˆ which generally comes quite close to the true value of θ, and dislike an estimate θˆ which generally misses the true value of θ by a large amount. How, are we going to make this precise and quantifiable? The general approach to this question involves the specification of a loss and a risk function, which we will introduce in a later section. For now, we describe a very common and even hugely popular criterion for evaluating a point estimator, the mean squared error. Definition 5.2. Let θ be a real valued parameter, and θˆ an estimate of θ. The mean squared error (MSE) of θˆ is defined as 2 MSE = MSE(θ, θˆ)= Eθ[(θˆ − θ) ], θ ∈ Θ. If the parameter θ is p-dimensional, θ = (θ1, ··· , θp), and θˆ = θˆ1, ··· , θˆp), then the mean squared error f θˆ is defined as p 2 2 MSE = MSE(θ, θˆ)= Eθ[||θˆ − θ|| ]= Eθ[(θˆi − θi) ], θ ∈ Θ. Xi=1 5.0.2 Bias and Variance It turns out that the mean squared error of an estimator has one component to do with systematic error of the estimator and a second component to do with random error of the estimator. If an estimator θˆ routinely overestimated the parameter θ, then usually we will have θˆ − θ > 0; we think of this as the estimator making a systematic error. Systematic 268 errors can also be made by routinely underestimating θ, We quantify the systematic error of an estimator by looking at Eθ(θˆ−θ)= Eθ(θˆ)−θ; this is called the bias of the estimator; we denote it as b(θ). If the bias of an estimator θˆ is always zero, b(θ) = 0 for all θ, then the estimator θˆ is called unbiased. On the other hand, the estimator θˆ may not make much systematic error, but still can just be unreliable because from one dataset to another, its accuracy may differ wildly. This is called random or fluctuation error, and we often quantify the random error by looking at the variance of the estimator, Varθ(θˆ). A pleasant property of the MSE of an estimator is that always the MSE neatly breaks into two components, one involving the bias, and the other involving the variance. You have to try to keep both of them small; large biases and large variances are both red signals. Here is a bias-variance decomposition result. Theorem 5.1. Let θ be a real valued parameter and θˆ an estimator with a finite variance under all θ. Then, 2 MSE(θ, θˆ) = Varθ(θˆ)+ b (θ), θ ∈ Θ. Proof; To prove this simple theorem, we recall the elementary probability fact that for any random variable U with a finitte variance, E(U 2) = Var(U)+[E(U)]2. Identifying U with θˆ − θ, 2 MSE(θ, θˆ) = Varθ(θˆ − θ)+[E(θˆ − θ)] 2 = Varθ(θˆ)+ b (θ). 5.0.3 Computing and Graphing MSE We will now see one introductory example. Example 5.2. (Estimating a Normal Mean and Variance) Suppose we have sample iid 2 observations X1, ··· , Xn ∼ N(µ, σ ), where −∞ <µ< ∞,σ > 0 are unknoown parame- ters. The parameter is two dimensional, θ = (µ, σ). First consider estimation of µ, and as an example, consider these two estimates: X¯ and nX¯ n+1 . We will calculate the MSE of each estimate, and make some comments. ¯ 1 n 1 ¯ Since E(X)= n sumi=1E(Xi)= n (nµ)= µ for any µ and σ, the bias of X for estimating µ is zero; Eθ(X¯) − µ = µ − µ = 0. In other words, X¯ is an unbiased estimate of µ. Therefore, the MSE of X¯ is just its variance, Var(X ) σ2 E[(X¯ − µ)2] = Var(X¯)= 1 = . n n Notice that the MSE of X¯ does not depend on µ; it only depends on σ2. nX¯ Now, we will find the MSE of the other estimate n+1 . For this, by our Theorem 3.1, the nX¯ MSE of n+1 is nX¯ nX¯ nX¯ E[( − µ)2] = Var( )+[E( − µ)]2 n + 1 n + 1 n + 1 269 Comparison of MSE of Estimates in Example 3.2 0.20 0.15 0.10 0.05 mu -4 -2 2 4 n n n σ2 1 = ( )2Var(X¯)+[ µ − µ]2 = ( )2 + ( )2µ2 n + 1 n + 1 n + 1 n n + 1 µ2 nσ2 = + . (n + 1)2 (n + 1)2 2 2 µ Notice that the MSE of this estimate does depend on both µ and σ ; (n+1)2 is the contri- nσ2 bution of the bias component in the MSE, and (n+1)2 is the contribution of the variance component. For purposes of comparison, we plot the MSE of both estimates, taking n = 10 and σ = 1; ¯ nX¯ the MSE of X is constant in µ, and the MSE of n+1 is a quadratic in µ. For µ near zero, nX¯ ¯ n+1 has a smaller MSE, but otherwise, X has a smaller MSE.
Recommended publications
  • Phase Transition Unbiased Estimation in High Dimensional Settings Arxiv
    Phase Transition Unbiased Estimation in High Dimensional Settings St´ephaneGuerrier, Mucyo Karemera, Samuel Orso & Maria-Pia Victoria-Feser Research Center for Statistics, University of Geneva Abstract An important challenge in statistical analysis concerns the control of the finite sample bias of estimators. This problem is magnified in high dimensional settings where the number of variables p diverge with the sample size n. However, it is difficult to establish whether an estimator θ^ of θ0 is unbiased and the asymptotic ^ order of E[θ] − θ0 is commonly used instead. We introduce a new property to assess the bias, called phase transition unbiasedness, which is weaker than unbiasedness but stronger than asymptotic results. An estimator satisfying this property is such ^ ∗ that E[θ] − θ0 2 = 0, for all n greater than a finite sample size n . We propose a phase transition unbiased estimator by matching an initial estimator computed on the sample and on simulated data. It is computed using an algorithm which is shown to converge exponentially fast. The initial estimator is not required to be consistent and thus may be conveniently chosen for computational efficiency or for other properties. We demonstrate the consistency and the limiting distribution of the estimator in high dimension. Finally, we develop new estimators for logistic regression models, with and without random effects, that enjoy additional properties such as robustness to data contamination and to the problem of separability. arXiv:1907.11541v3 [math.ST] 1 Nov 2019 Keywords: Finite sample bias, Iterative bootstrap, Two-step estimators, Indirect inference, Robust estimation, Logistic regression. 1 1. Introduction An important challenge in statistical analysis concerns the control of the (finite sample) bias of estimators.
    [Show full text]
  • Stable Components in the Parameter Plane of Transcendental Functions of Finite Type
    Stable components in the parameter plane of transcendental functions of finite type N´uriaFagella∗ Dept. de Matem`atiquesi Inform`atica Univ. de Barcelona, Gran Via 585 Barcelona Graduate School of Mathematics (BGSMath) 08007 Barcelona, Spain [email protected] Linda Keen† CUNY Graduate Center 365 Fifth Avenue, New York NY, NY 10016 [email protected], [email protected] November 11, 2020 Abstract We study the parameter planes of certain one-dimensional, dynamically-defined slices of holomorphic families of entire and meromorphic transcendental maps of finite type. Our planes are defined by constraining the orbits of all but one of the singular values, and leaving free one asymptotic value. We study the structure of the regions of parameters, which we call shell components, for which the free asymptotic value tends to an attracting cycle of non-constant multiplier. The exponential and the tangent families are examples that have been studied in detail, and the hyperbolic components in those parameter planes are shell components. Our results apply to slices of both entire and meromorphic maps. We prove that shell components are simply connected, have a locally connected boundary and have no center, i.e., no parameter value for which the cycle is superattracting. Instead, there is a unique parameter in the boundary, the virtual center, which plays the same role. For entire slices, the virtual center is always at infinity, while for meromorphic ones it maybe finite or infinite. In the dynamical plane we prove, among other results, that the basins of attraction which contain only one asymptotic value and no critical points are simply connected.
    [Show full text]
  • Bias, Mean-Square Error, Relative Efficiency
    3 Evaluating the Goodness of an Estimator: Bias, Mean-Square Error, Relative Efficiency Consider a population parameter ✓ for which estimation is desired. For ex- ample, ✓ could be the population mean (traditionally called µ) or the popu- lation variance (traditionally called σ2). Or it might be some other parame- ter of interest such as the population median, population mode, population standard deviation, population minimum, population maximum, population range, population kurtosis, or population skewness. As previously mentioned, we will regard parameters as numerical charac- teristics of the population of interest; as such, a parameter will be a fixed number, albeit unknown. In Stat 252, we will assume that our population has a distribution whose density function depends on the parameter of interest. Most of the examples that we will consider in Stat 252 will involve continuous distributions. Definition 3.1. An estimator ✓ˆ is a statistic (that is, it is a random variable) which after the experiment has been conducted and the data collected will be used to estimate ✓. Since it is true that any statistic can be an estimator, you might ask why we introduce yet another word into our statistical vocabulary. Well, the answer is quite simple, really. When we use the word estimator to describe a particular statistic, we already have a statistical estimation problem in mind. For example, if ✓ is the population mean, then a natural estimator of ✓ is the sample mean. If ✓ is the population variance, then a natural estimator of ✓ is the sample variance. More specifically, suppose that Y1,...,Yn are a random sample from a population whose distribution depends on the parameter ✓.The following estimators occur frequently enough in practice that they have special notations.
    [Show full text]
  • Backpropagation TA: Yi Wen
    Backpropagation TA: Yi Wen April 17, 2020 CS231n Discussion Section Slides credits: Barak Oshri, Vincent Chen, Nish Khandwala, Yi Wen Agenda ● Motivation ● Backprop Tips & Tricks ● Matrix calculus primer Agenda ● Motivation ● Backprop Tips & Tricks ● Matrix calculus primer Motivation Recall: Optimization objective is minimize loss Motivation Recall: Optimization objective is minimize loss Goal: how should we tweak the parameters to decrease the loss? Agenda ● Motivation ● Backprop Tips & Tricks ● Matrix calculus primer A Simple Example Loss Goal: Tweak the parameters to minimize loss => minimize a multivariable function in parameter space A Simple Example => minimize a multivariable function Plotted on WolframAlpha Approach #1: Random Search Intuition: the step we take in the domain of function Approach #2: Numerical Gradient Intuition: rate of change of a function with respect to a variable surrounding a small region Approach #2: Numerical Gradient Intuition: rate of change of a function with respect to a variable surrounding a small region Finite Differences: Approach #3: Analytical Gradient Recall: partial derivative by limit definition Approach #3: Analytical Gradient Recall: chain rule Approach #3: Analytical Gradient Recall: chain rule E.g. Approach #3: Analytical Gradient Recall: chain rule E.g. Approach #3: Analytical Gradient Recall: chain rule Intuition: upstream gradient values propagate backwards -- we can reuse them! Gradient “direction and rate of fastest increase” Numerical Gradient vs Analytical Gradient What about Autograd?
    [Show full text]
  • Decomposing the Parameter Space of Biological Networks Via a Numerical Discriminant Approach
    Decomposing the parameter space of biological networks via a numerical discriminant approach Heather A. Harrington1, Dhagash Mehta2, Helen M. Byrne1, and Jonathan D. Hauenstein2 1 Mathematical Institute, The University of Oxford, Oxford OX2 6GG, UK {harrington,helen.byrne}@maths.ox.ac.uk, www.maths.ox.ac.uk/people/{heather.harrington,helen.byrne} 2 Department of Applied and Computational Mathematics and Statistics, University of Notre Dame, Notre Dame IN 46556, USA {dmehta,hauenstein}@nd.edu, www.nd.edu/~{dmehta,jhauenst} Abstract. Many systems in biology (as well as other physical and en- gineering systems) can be described by systems of ordinary dierential equation containing large numbers of parameters. When studying the dynamic behavior of these large, nonlinear systems, it is useful to iden- tify and characterize the steady-state solutions as the model parameters vary, a technically challenging problem in a high-dimensional parameter landscape. Rather than simply determining the number and stability of steady-states at distinct points in parameter space, we decompose the parameter space into nitely many regions, the number and structure of the steady-state solutions being consistent within each distinct region. From a computational algebraic viewpoint, the boundary of these re- gions is contained in the discriminant locus. We develop global and local numerical algorithms for constructing the discriminant locus and classi- fying the parameter landscape. We showcase our numerical approaches by applying them to molecular and cell-network models. Keywords: parameter landscape · numerical algebraic geometry · dis- criminant locus · cellular networks. 1 Introduction The dynamic behavior of many biophysical systems can be mathematically mod- eled with systems of dierential equations that describe how the state variables interact and evolve over time.
    [Show full text]
  • Weak Instruments and Finite-Sample Bias
    7 Weak instruments and finite-sample bias In this chapter, we consider the effect of weak instruments on instrumen- tal variable (IV) analyses. Weak instruments, which were introduced in Sec- tion 4.5.2, are those that do not explain a large proportion of the variation in the exposure, and so the statistical association between the IV and the expo- sure is not strong. This is of particular relevance in Mendelian randomization studies since the associations of genetic variants with exposures of interest are often weak. This chapter focuses on the impact of weak instruments on the bias and coverage of IV estimates. 7.1 Introduction Although IV techniques can be used to give asymptotically unbiased estimates of causal effects in the presence of confounding, these estimates suffer from bias when evaluated in finite samples [Nelson and Startz, 1990]. A weak instrument (or a weak IV) is still a valid IV, in that it satisfies the IV assumptions, and es- timates using the IV with an infinite sample size will be unbiased; but for any finite sample size, the average value of the IV estimator will be biased. This bias, known as weak instrument bias, is towards the observational confounded estimate. Its magnitude depends on the strength of association between the IV and the exposure, which is measured by the F statistic in the regression of the exposure on the IV [Bound et al., 1995]. In this chapter, we assume the context of ‘one-sample’ Mendelian randomization, in which evidence on the genetic variant, exposure, and outcome are taken on the same set of indi- viduals, rather than subsample (Section 8.5.2) or two-sample (Section 9.8.2) Mendelian randomization, in which genetic associations with the exposure and outcome are estimated in different sets of individuals (overlapping sets in subsample, non-overlapping sets in two-sample Mendelian randomization).
    [Show full text]
  • The Emergence of Gravitational Wave Science: 100 Years of Development of Mathematical Theory, Detectors, Numerical Algorithms, and Data Analysis Tools
    BULLETIN (New Series) OF THE AMERICAN MATHEMATICAL SOCIETY Volume 53, Number 4, October 2016, Pages 513–554 http://dx.doi.org/10.1090/bull/1544 Article electronically published on August 2, 2016 THE EMERGENCE OF GRAVITATIONAL WAVE SCIENCE: 100 YEARS OF DEVELOPMENT OF MATHEMATICAL THEORY, DETECTORS, NUMERICAL ALGORITHMS, AND DATA ANALYSIS TOOLS MICHAEL HOLST, OLIVIER SARBACH, MANUEL TIGLIO, AND MICHELE VALLISNERI In memory of Sergio Dain Abstract. On September 14, 2015, the newly upgraded Laser Interferometer Gravitational-wave Observatory (LIGO) recorded a loud gravitational-wave (GW) signal, emitted a billion light-years away by a coalescing binary of two stellar-mass black holes. The detection was announced in February 2016, in time for the hundredth anniversary of Einstein’s prediction of GWs within the theory of general relativity (GR). The signal represents the first direct detec- tion of GWs, the first observation of a black-hole binary, and the first test of GR in its strong-field, high-velocity, nonlinear regime. In the remainder of its first observing run, LIGO observed two more signals from black-hole bina- ries, one moderately loud, another at the boundary of statistical significance. The detections mark the end of a decades-long quest and the beginning of GW astronomy: finally, we are able to probe the unseen, electromagnetically dark Universe by listening to it. In this article, we present a short historical overview of GW science: this young discipline combines GR, arguably the crowning achievement of classical physics, with record-setting, ultra-low-noise laser interferometry, and with some of the most powerful developments in the theory of differential geometry, partial differential equations, high-performance computation, numerical analysis, signal processing, statistical inference, and data science.
    [Show full text]
  • STAT 830 the Basics of Nonparametric Models The
    STAT 830 The basics of nonparametric models The Empirical Distribution Function { EDF The most common interpretation of probability is that the probability of an event is the long run relative frequency of that event when the basic experiment is repeated over and over independently. So, for instance, if X is a random variable then P (X ≤ x) should be the fraction of X values which turn out to be no more than x in a long sequence of trials. In general an empirical probability or expected value is just such a fraction or average computed from the data. To make this precise, suppose we have a sample X1;:::;Xn of iid real valued random variables. Then we make the following definitions: Definition: The empirical distribution function, or EDF, is n 1 X F^ (x) = 1(X ≤ x): n n i i=1 This is a cumulative distribution function. It is an estimate of F , the cdf of the Xs. People also speak of the empirical distribution of the sample: n 1 X P^(A) = 1(X 2 A) n i i=1 ^ This is the probability distribution whose cdf is Fn. ^ Now we consider the qualities of Fn as an estimate, the standard error of the estimate, the estimated standard error, confidence intervals, simultaneous confidence intervals and so on. To begin with we describe the best known summaries of the quality of an estimator: bias, variance, mean squared error and root mean squared error. Bias, variance, MSE and RMSE There are many ways to judge the quality of estimates of a parameter φ; all of them focus on the distribution of the estimation error φ^−φ.
    [Show full text]
  • Bias in Parametric Estimation: Reduction and Useful Side-Effects
    Bias in parametric estimation: reduction and useful side-effects Ioannis Kosmidis Department of Statistical Science, University College London August 30, 2018 Abstract The bias of an estimator is defined as the difference of its expected value from the parameter to be estimated, where the expectation is with respect to the model. Loosely speaking, small bias reflects the desire that if an experiment is repeated in- definitely then the average of all the resultant estimates will be close to the parameter value that is estimated. The current paper is a review of the still-expanding repository of methods that have been developed to reduce bias in the estimation of parametric models. The review provides a unifying framework where all those methods are seen as attempts to approximate the solution of a simple estimating equation. Of partic- ular focus is the maximum likelihood estimator, which despite being asymptotically unbiased under the usual regularity conditions, has finite-sample bias that can re- sult in significant loss of performance of standard inferential procedures. An informal comparison of the methods is made revealing some useful practical side-effects in the estimation of popular models in practice including: i) shrinkage of the estimators in binomial and multinomial regression models that guarantees finiteness even in cases of data separation where the maximum likelihood estimator is infinite, and ii) inferential benefits for models that require the estimation of dispersion or precision parameters. Keywords: jackknife/bootstrap, indirect inference, penalized likelihood, infinite es- timates, separation in models with categorical responses 1 Impact of bias in estimation By its definition, bias necessarily depends on how the model is written in terms of its parameters and this dependence makes it not a strong statistical principle in terms of evaluating the performance of estimators; for example, unbiasedness of the familiar sample variance S2 as an estimator of σ2 does not deliver an unbiased estimator of σ itself.
    [Show full text]
  • Nearly Weighted Risk Minimal Unbiased Estimation✩ ∗ Ulrich K
    Journal of Econometrics 209 (2019) 18–34 Contents lists available at ScienceDirect Journal of Econometrics journal homepage: www.elsevier.com/locate/jeconom Nearly weighted risk minimal unbiased estimationI ∗ Ulrich K. Müller a, , Yulong Wang b a Economics Department, Princeton University, United States b Economics Department, Syracuse University, United States article info a b s t r a c t Article history: Consider a small-sample parametric estimation problem, such as the estimation of the Received 26 July 2017 coefficient in a Gaussian AR(1). We develop a numerical algorithm that determines an Received in revised form 7 August 2018 estimator that is nearly (mean or median) unbiased, and among all such estimators, comes Accepted 27 November 2018 close to minimizing a weighted average risk criterion. We also apply our generic approach Available online 18 December 2018 to the median unbiased estimation of the degree of time variation in a Gaussian local-level JEL classification: model, and to a quantile unbiased point forecast for a Gaussian AR(1) process. C13, C22 ' 2018 Elsevier B.V. All rights reserved. Keywords: Mean bias Median bias Autoregression Quantile unbiased forecast 1. Introduction Competing estimators are typically evaluated by their bias and risk properties, such as their mean bias and mean-squared error, or their median bias. Often estimators have no known small sample optimality. What is more, if the estimation problem does not reduce to a Gaussian shift experiment even asymptotically, then in many cases, no optimality
    [Show full text]
  • Investigations of Structures in the Parameter Space of Three-Dimensional Turing-Like Patterns Martin Skrodzki, Ulrich Reitebuch, Eric Zimmermann
    Investigations of structures in the parameter space of three-dimensional Turing-like patterns Martin Skrodzki, Ulrich Reitebuch, Eric Zimmermann To cite this version: Martin Skrodzki, Ulrich Reitebuch, Eric Zimmermann. Investigations of structures in the parameter space of three-dimensional Turing-like patterns. AUTOMATA2021, Jul 2021, Marseille, France. hal- 03270664 HAL Id: hal-03270664 https://hal.archives-ouvertes.fr/hal-03270664 Submitted on 25 Jun 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Investigations of structures in the parameter space of three-dimensional Turing-like patterns∗ Martin Skrodzki,† Computer Graphics and Visualization, TU Delft [email protected] Ulrich Reitebuch, Institute of Mathematics, FU Berlin [email protected] Eric Zimmermann, Institute of Mathematics, FU Berlin [email protected] Abstract In this paper, we are interested in classifying the different arising (topological) structures of three-dimensional Turing-like patterns. By providing examples for the different structures, we confirm a conjec- ture regarding these structures within the setup of three-dimensional Turing-like pattern. Furthermore, we investigate how these structures are distributed in the parameter space of the discrete model. We found two-fold versions of so-called \zero-" and \one-dimensional" structures as well as \two-dimensional" structures and use our experimental find- ings to formulate several conjectures for three-dimensional Turing-like patterns and higher-dimensional cases.
    [Show full text]
  • Chapter 4 Parameter Estimation
    Chapter 4 Parameter Estimation Thus far we have concerned ourselves primarily with probability theory: what events may occur with what probabilities, given a model family and choices for the parameters. This is useful only in the case where we know the precise model family and parameter values for the situation of interest. But this is the exception, not the rule, for both scientific inquiry and human learning & inference. Most of the time, we are in the situation of processing data whose generative source we are uncertain about. In Chapter 2 we briefly covered elemen- tary density estimation, using relative-frequency estimation, histograms and kernel density estimation. In this chapter we delve more deeply into the theory of probability density es- timation, focusing on inference within parametric families of probability distributions (see discussion in Section 2.11.2). We start with some important properties of estimators, then turn to basic frequentist parameter estimation (maximum-likelihood estimation and correc- tions for bias), and finally basic Bayesian parameter estimation. 4.1 Introduction Consider the situation of the first exposure of a native speaker of American English to an English variety with which she has no experience (e.g., Singaporean English), and the problem of inferring the probability of use of active versus passive voice in this variety with a simple transitive verb such as hit: (1) The ball hit the window. (Active) (2) The window was hit by the ball. (Passive) There is ample evidence that this probability is contingent
    [Show full text]