Random Fields, Fall 2014 1

Total Page:16

File Type:pdf, Size:1020Kb

Random Fields, Fall 2014 1 Random fields, Fall 2014 1 fMRI brain scan Ocean waves Nobel prize 2996 to The oceans cover 72% of John C. Mather the earth’s surface. George F. Smoot Essential for life on earth, “for discovery of the and huge economic blackbody form and importance through anisotropy of the cosmic fishing, transportation, microwave background oil and gas extraction radiation" PET brain scan TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A The course • Kolmogorov existence theorem, separable processes, measurable processes • Stationarity and isotropy • Orthogonal and spectral representations • Geometry • Exceedance sets • Rice formula • Slepian models Literature An unfinished manuscript “Applications of RANDOM FIELDS AND GEOMETRY: Foundations and Case Studies” by Robert Adler, Jonathan Taylor, and Keith Worsley. Complementary literature: “Level sets and extrema of random processes and fields” by Jean- Marc Azais and Mario Wschebor, Wiley, 2009 “Asymptotic Methods in the Theory of Gaussian Processes” by Vladimir Piterbarg, American Mathematical Society, ser. Translations of Mathematical Monographs, Vol. 148, 1995 “Random fields and Geometry” by Robert Adler and Jonathan Taylor, Springer 2007 Slides for ATW ch 2, p 24-39 Exercises: 2.8.1, 2.8.2, 2.8.3, 2.8.4, 2.8.5, 2.8.6 + excercises in slides Stochastic convergence (assumed known) 푎.푠. Almost sure convergence: 푋푛 X 퐿2 Mean square convergence: 푋푛 X 푃 Convergence in probability: 푋푛 X 푑 푑 푑 Convergence in distribution: 푋푛 X, 퐹푛 F, 푃푛 P, … 푎.푠. 푃 • 푋푛 X ⇒ 푋푛 푋 퐿2 푃 • 푋푛 X ⇒ 푋푛 푋 푃 퐿2 • 푋푛 푋 plus uniform integrability ⇒ 푋푛 X 푃 푎.푠. • 푋푛 푋 ⇒ there is a subsequence {푛푘} with 푋푛푘 X 푑 • The random variables don’t really mean anything for In particular, the 푋푛 and 푋 don’t need to be defined on the same probability space, and don’t need to have a simultaneous distribution Random field 푇 parameter space. In this course 푇 is 푹푁 some 푁 ≥ 1 or a subset (e.g. box or sphere or surface of sphere) of 푹푁 푹푑 value space An (푁, 푑) random field is a collection (or family) of random variables 푓푡; 푡 ∈ 푇 where 푇 is a set of dimension 푁 and the 푓푡 (or 푓(푡)) take values in 푹푑 푇 Or, a random function with values in 푹푑 푇 Or, a probablility measure on 푹푑 A realisation (or sample function, or sample path, or sample field, or observation, or trajectory, or …) is the function 푑 푓푡 휔 : 푇 푹 푡 ft(휔) for 휔 fixed. Two examples below: Microscopy image of Thresholded Gaussian tablet coating field Terminology random variable stochastic variable random element random vector stochastic element stochastic vector random field random process stochastic field stochastic process Finite dimensional distributions The finite-dimensional distribution functions of an (푁, 푑) random field {푓푡} are defined as 퐹푡1,…,푡푛 풙1, … , 풙푛 = 푃(푓푡1 ≤ 풙1, … , 푓푡푛 ≤ 풙n) and the family of finite-dimensional distribution functions is the set 푑 {퐹푡1,…,푡푛 풙1, … , 풙푛 ; 푡1, … , 푡푛 ∈ 푇, 풙1, … , 풙푛 ∈ 푅 , 푛 ≥ 1} This family has the following obvious properties: Symmetry: it is not changed under a simultaneous permutation of 푡1, … , 푡푛 and 풙1, … , 풙푛 Consistency: 퐹푡1,…,푡푛 풙1, … , 풙푛−1, ∞ = 퐹푡1,…,푡푛−1 풙1, … , 풙푛−1 Example of symmetry: 퐹푡1,푡2 풙1, 풙2 = 퐹푡2,푡1 풙2, 풙1 Example of consistency: Marginal distributions may be obtained from bivariate distributions, 퐹푡 풙 = 퐹푡,푠 풙, ∞ Three sample paths of a 1,1 random field. 퐹2,5,8 푥1, 푥2, 푥3 is the probability to obtain a sample path which passes through all three vertical lines (in a more general theory one instead of finite-dimensional distributions uses probabilities of cylindersets, 푃(푓푡1 ∈ 푩1, … , 푓푡푛 ∈ 푩n) ) (Daniell-)Kolmogorov extension theorem To any symmetric and consistent family of finite-dimensional distributions 푑 {퐹푡1,…,푡푛 풙1, … , 풙푛 ; 푡1, … , 푡푛 ∈ 푇, 풙1, … , 풙푛 ∈ 푅 , 푛 ≥ 1} there exists a probability triple (Ω, , 푃) and an 푁, 푑 B random field {푓푡; 푡 ∈ 푇 } which has these finite-dimensional distributions In the proof one takes 푑 푇 푑 푇 Ω = 푹 , = (푹 ) , 푇 and 푃 as the measure on (푹푑) which is uniquely determined by the finite-dimensional distribution. Thus an element of Ω is a function 푓: 푇 푅푑 which maps a point 푡 ∈ 푇 to the value 푓(푡). The field is defined as {푓푡 휔 = 푓 푡 ; 푡 ∈ 푇} Limitations of Kolmogorov’s theorem Many interesting sets, such as the set 퐶 = {휔; 푓푡 휔 is a continous function of 푡} 푇 do not belong to = (푹푑) , and hence, in Kolmogorov’s construction , the probababilityB of such events is not defined. One important way around this problem is to make a direct construction of the field on some other probability space (Ω, , 푃) where the interesting sets belong to , say 퐶 ∈ , so that their probabilities, say 푃(퐶), is well defined. And then, more fields are obtained as functions of the already constructed field! Modifications A field 푔푡; 푡 ∈ 푇 is a modification of the field 푓푡; 푡 ∈ 푇 if 푃 푔푡 = 푓푡 = 1, ∀푡 ∈ 푇 It is obvious (!) that 푔푡 has the same finite dimensional distributions as 푓푡. B The other common way to circumvent the limitation is to construct, on Kolmogorov’s (Ω, , 푃) a modification of 푓푡 which has the desired properties, say continuity. Whether this is possible or not (of course) dependes on which finite-dimensional distributions one is interested in. E.g. if they correspond to a Browninan motion it is possible, if they correspond to a Poisson process, it isn’t. Doob’s separability A field 푓푡; 푡 ∈ 푇 is separable if there is a countable subset 푇 푆 ∈ 푇 and a null set Λ ∈ (푹푑) such that for every closed set 퐵 ∈ 푹푑 and open set 퐼 ∈ 푇 it holds that 푓푡 휔 ∈ 퐵, ∀푡 ∈ 푆 ∩ 퐼 ⇒ 휔 ∈ ΛB or 푓푡 휔 ∈ 퐵, ∀푡 ∈ 퐼 A separable modification of a field always exists (at least for 푁 = 푑 = 1? ), and it can be seen that e.g. if a continuous modification of a field exists, then the separable modification is continuous. Example of modification: Ω = 0,1 , is the Borel sets on [0, 1], 푃 is Lebesgue measure, 푓푡 휔 = 0, ∀푡, 휔 and 0 if 푡 ≠ 휔 푔 휔 = 푡 1 if 푡 = 휔 Measurable fields A field 푓푡; 푡 ∈ 푇 is measurable if for almost all 휔 the sample path (function) 푑 푓 . 휔 : 푇 푅 B 푡 푓푡(휔) is (푅푑)-measurable (holds e.g. if the field is a.s. continuous). It then follows that the function of two variables 푓푡(휔) is measurable with respect to the product sigma-algebra × (푅푑), and one can then define integrals like 푇 ℎ 푓푡 푑푡 and use Fubini’s theorem for calculations like 퐸 ℎ 푓푡 푑푡 = 퐸(ℎ 푓푡 )푑푡 푇 푇 (above we have assumed that and (푅푑) are complete) ATW basically say that it is nice if one has seen the concepts of Kolmogorov extension, modification, and Doob separability, but that this has been taken care of once and for all by Kolmogorov, Doob and others, and that we shouldn’t worry about it any more in this course. And this is right (I hope). However, things are different for the theory of ”Empirical Processes”, the so far most efficient and high-tech tool to find asymptotic distributions of statistical estimators. In this theory, such ”measureability problems” pose important techical problems, and has formed much of the entire theory. Empirical process theory is closely related to the metods used to prove continuity and differentiability in this course. Gaussian fields (ATW p. 25-28) A random vector 퐗 = 푋1, … , 푋푑 has a multivariate Gaussian distribution iff one of the following conditions hold: 푑 • 훼, 푥 ≜ 푖 훼푖푋푖 has a univariate normal distribution for all 훼 ∈ 푅푑. • There exist a vector 풎 ∈ 푹푑 and a non-negative definite matrix 퐶 such that for all 휽 ∈ 푹푑 1 휙 휽 = 퐸 푒푖휽푋 = e푖휃풎−2휽퐶휽´ If 퐶 is positive definite and 푋 has the probability density 1 1 푒−2 풙−풎 퐶 풙−풎 ´ 2휋 푑 퐶 1/2 then 푋 is Gaussian. 풅 Here 푚 = 퐸 푿 and 퐶 = 퐶표푣 푿 . Similarly if the 푿풊 ∈ 푹 We write 퐗~푁푑 풎, 퐶 if 퐗 has a d-variate Gaussian distribution with mean 푚 and covariance matrix 퐶. Excercises (the first is (2.2.5), the second Exercise 2.8.2): (i) if 퐗~푁푑 풎, 퐶 and 퐴 is a 푑 × 푑 matrix, then 퐗퐴~푁푑 풎퐴, 퐴´퐶퐴 1 2 1 2 (ii) If 푿 = 푿 , 푿 with 푿 = 푋1, … , 푋푛 , 푿 = 푋푛+1, … , 푋푑 , with mean vectors 푚1 and 푚2 and covariance matrix 퐶1,1 퐶1,2 1 퐶 = , then the conditional distribution of 푿 given 퐶2,1 퐶2,2 2 푿 is n-variate normal with mean 1 2 2 −1 풎1|2 = 풎 + (푿 −풎 )퐶2,2 퐶2,1 and covariance matrix −1 퐶1|2 = 퐶1,1 − 퐶1,2퐶2,2 퐶2,1 A Gaussian random field is hence, by the Kolmogorov theorem, determined by its means and covariances Conversely, it also follows from the Kolmogorov theorem that given a function 푚: 푇 푹 and a non-negative definite function 퐶: 푇 × 푇 푹 there exist an 푁, 1 Gaussian random field which has 풎 as mean function and 퐶 as covariance function. 푁, 푑 Gaussian random fields for 푑 > 1 are the same, one just has to use more general notation. Gaussian related fields (ATW p. 28-30) An (푁, 푑) Gaussian related field 푓(푡); 푡 ∈ 푇 is defined from a (푁, 푘) Gaussian field 푔(푡); 푡 ∈ 푇 using a function 푘 푑 퐹: 푅 푅 by the formula 푓 푡 = 퐹 푔 푡 = 퐹 g1 t , … , 푔푘 푡 .
Recommended publications
  • Randomized Algorithms
    Chapter 9 Randomized Algorithms randomized The theme of this chapter is madendrizo algorithms. These are algorithms that make use of randomness in their computation. You might know of quicksort, which is efficient on average when it uses a random pivot, but can be bad for any pivot that is selected without randomness. Analyzing randomized algorithms can be difficult, so you might wonder why randomiza- tion in algorithms is so important and worth the extra effort. Well it turns out that for certain problems randomized algorithms are simpler or faster than algorithms that do not use random- ness. The problem of primality testing (PT), which is to determine if an integer is prime, is a good example. In the late 70s Miller and Rabin developed a famous and simple random- ized algorithm for the problem that only requires polynomial work. For over 20 years it was not known whether the problem could be solved in polynomial work without randomization. Eventually a polynomial time algorithm was developed, but it is much more complicated and computationally more costly than the randomized version. Hence in practice everyone still uses the randomized version. There are many other problems in which a randomized solution is simpler or cheaper than the best non-randomized solution. In this chapter, after covering the prerequisite background, we will consider some such problems. The first we will consider is the following simple prob- lem: Question: How many comparisons do we need to find the top two largest numbers in a sequence of n distinct numbers? Without the help of randomization, there is a trivial algorithm for finding the top two largest numbers in a sequence that requires about 2n − 3 comparisons.
    [Show full text]
  • MA3K0 - High-Dimensional Probability
    MA3K0 - High-Dimensional Probability Lecture Notes Stefan Adams i 2020, update 06.05.2020 Notes are in final version - proofreading not completed yet! Typos and errors will be updated on a regular basis Contents 1 Prelimaries on Probability Theory 1 1.1 Random variables . .1 1.2 Classical Inequalities . .4 1.3 Limit Theorems . .6 2 Concentration inequalities for independent random variables 8 2.1 Why concentration inequalities . .8 2.2 Hoeffding’s Inequality . 10 2.3 Chernoff’s Inequality . 13 2.4 Sub-Gaussian random variables . 15 2.5 Sub-Exponential random variables . 22 3 Random vectors in High Dimensions 27 3.1 Concentration of the Euclidean norm . 27 3.2 Covariance matrices and Principal Component Analysis (PCA) . 30 3.3 Examples of High-Dimensional distributions . 32 3.4 Sub-Gaussian random variables in higher dimensions . 34 3.5 Application: Grothendieck’s inequality . 36 4 Random Matrices 36 4.1 Geometrics concepts . 36 4.2 Concentration of the operator norm of random matrices . 38 4.3 Application: Community Detection in Networks . 41 4.4 Application: Covariance Estimation and Clustering . 41 5 Concentration of measure - general case 43 5.1 Concentration by entropic techniques . 43 5.2 Concentration via Isoperimetric Inequalities . 50 5.3 Some matrix calculus and covariance estimation . 54 6 Basic tools in high-dimensional probability 57 6.1 Decoupling . 57 6.2 Concentration for Anisotropic random vectors . 62 6.3 Symmetrisation . 63 7 Random Processes 64 ii Preface Introduction We discuss an elegant argument that showcases the
    [Show full text]
  • HIGHER MOMENTS of BANACH SPACE VALUED RANDOM VARIABLES 1. Introduction Let X Be a Random Variable with Values in a Banach Space
    HIGHER MOMENTS OF BANACH SPACE VALUED RANDOM VARIABLES SVANTE JANSON AND STEN KAIJSER Abstract. We define the k:th moment of a Banach space valued ran- dom variable as the expectation of its k:th tensor power; thus the mo- ment (if it exists) is an element of a tensor power of the original Banach space. We study both the projective and injective tensor products, and their relation. Moreover, in order to be general and flexible, we study three different types of expectations: Bochner integrals, Pettis integrals and Dunford integrals. One of the problems studied is whether two random variables with the same injective moments (of a given order) necessarily have the same projective moments; this is of interest in applications. We show that this holds if the Banach space has the approximation property, but not in general. Several sections are devoted to results in special Banach spaces, in- cluding Hilbert spaces, CpKq and Dr0; 1s. The latter space is non- separable, which complicates the arguments, and we prove various pre- liminary results on e.g. measurability in Dr0; 1s that we need. One of the main motivations of this paper is the application to Zolotarev metrics and their use in the contraction method. This is sketched in an appendix. 1. Introduction Let X be a random variable with values in a Banach space B. To avoid measurability problems, we assume for most of this section for simplicity that B is separable and X Borel measurable; see Section 3 for measurability in the general case. Moreover, for definiteness, we consider real Banach spaces only; the complex case is similar.
    [Show full text]
  • Control Policy with Autocorrelated Noise in Reinforcement Learning for Robotics
    International Journal of Machine Learning and Computing, Vol. 5, No. 2, April 2015 Control Policy with Autocorrelated Noise in Reinforcement Learning for Robotics Paweł Wawrzyński primitives by means of RL algorithms based on MDP Abstract—Direct application of reinforcement learning in framework, but it does not alleviate the problem of robot robotics rises the issue of discontinuity of control signal. jerking. The current paper is intended to fill this gap. Consecutive actions are selected independently on random, In this paper, control policy is introduced that may undergo which often makes them excessively far from one another. Such reinforcement learning and have the following properties: control is hardly ever appropriate in robots, it may even lead to their destruction. This paper considers a control policy in which It is applicable to robot control optimization. consecutive actions are modified by autocorrelated noise. That It does not lead to robot jerking. policy generally solves the aforementioned problems and it is readily applicable in robots. In the experimental study it is It can be optimized by any RL algorithm that is designed applied to three robotic learning control tasks: Cart-Pole to optimize a classical stochastic control policy. SwingUp, Half-Cheetah, and a walking humanoid. The policy introduced here is based on a deterministic transformation of state combined with a random element in Index Terms—Machine learning, reinforcement learning, actorcritics, robotics. the form of a specific stochastic process, namely the moving average. The paper is organized as follows. In Section II the I. INTRODUCTION problem of our interest is defined. Section III presents the main contribution of this paper i.e., a stochastic control policy Reinforcement learning (RL) addresses the problem of an that prevents robot jerking while learning.
    [Show full text]
  • CONVERGENCE RATES of MARKOV CHAINS 1. Orientation 1
    CONVERGENCE RATES OF MARKOV CHAINS STEVEN P. LALLEY CONTENTS 1. Orientation 1 1.1. Example 1: Random-to-Top Card Shuffling 1 1.2. Example 2: The Ehrenfest Urn Model of Diffusion 2 1.3. Example 3: Metropolis-Hastings Algorithm 4 1.4. Reversible Markov Chains 5 1.5. Exercises: Reversiblity, Symmetries and Stationary Distributions 8 2. Coupling 9 2.1. Coupling and Total Variation Distance 9 2.2. Coupling Constructions and Convergence of Markov Chains 10 2.3. Couplings for the Ehrenfest Urn and Random-to-Top Shuffling 12 2.4. The Coupon Collector’s Problem 13 2.5. Exercises 15 2.6. Convergence Rates for the Ehrenfest Urn and Random-to-Top 16 2.7. Exercises 17 3. Spectral Analysis 18 3.1. Transition Kernel of a Reversible Markov Chain 18 3.2. Spectrum of the Ehrenfest random walk 21 3.3. Rate of convergence of the Ehrenfest random walk 23 1. ORIENTATION Finite-state Markov chains have stationary distributions, and irreducible, aperiodic, finite- state Markov chains have unique stationary distributions. Furthermore, for any such chain the n−step transition probabilities converge to the stationary distribution. In various ap- plications – especially in Markov chain Monte Carlo, where one runs a Markov chain on a computer to simulate a random draw from the stationary distribution – it is desirable to know how many steps it takes for the n−step transition probabilities to become close to the stationary probabilities. These notes will introduce several of the most basic and important techniques for studying this problem, coupling and spectral analysis. We will focus on two Markov chains for which the convergence rate is of particular interest: (1) the random-to-top shuffling model and (2) the Ehrenfest urn model.
    [Show full text]
  • Multivariate Scale-Mixed Stable Distributions and Related Limit
    Multivariate scale-mixed stable distributions and related limit theorems∗ Victor Korolev,† Alexander Zeifman‡ Abstract In the paper, multivariate probability distributions are considered that are representable as scale mixtures of multivariate elliptically contoured stable distributions. It is demonstrated that these distributions form a special subclass of scale mixtures of multivariate elliptically contoured normal distributions. Some properties of these distributions are discussed. Main attention is paid to the representations of the corresponding random vectors as products of independent random variables. In these products, relations are traced of the distributions of the involved terms with popular probability distributions. As examples of distributions of the class of scale mixtures of multivariate elliptically contoured stable distributions, multivariate generalized Linnik distributions are considered in detail. Their relations with multivariate ‘ordinary’ Linnik distributions, multivariate normal, stable and Laplace laws as well as with univariate Mittag-Leffler and generalized Mittag-Leffler distributions are discussed. Limit theorems are proved presenting necessary and sufficient conditions for the convergence of the distributions of random sequences with independent random indices (including sums of a random number of random vectors and multivariate statistics constructed from samples with random sizes) to scale mixtures of multivariate elliptically contoured stable distributions. The property of scale- mixed multivariate stable
    [Show full text]
  • Random Polynomials Over Finite Fields
    RANDOM POLYNOMIALS OVER FINITE FIELDS by Andrew Sharkey A thesis submitted to the Faculty of Science at the University of Glasgow for the degree of . Doctor of Philosophy June 22, 1999 ©Andrew Sharkey 1999 ProQuest Number: 13834241 All rights reserved INFORMATION TO ALL USERS The quality of this reproduction is dependent upon the quality of the copy submitted. In the unlikely event that the author did not send a complete manuscript and there are missing pages, these will be noted. Also, if material had to be removed, a note will indicate the deletion. uest ProQuest 13834241 Published by ProQuest LLC(2019). Copyright of the Dissertation is held by the Author. All rights reserved. This work is protected against unauthorized copying under Title 17, United States Code Microform Edition © ProQuest LLC. ProQuest LLC. 789 East Eisenhower Parkway P.O. Box 1346 Ann Arbor, Ml 4 8 1 0 6 - 1346 m Questa tesi e dedicata ad Alessandro Paluello (1975-1997). Preface This thesis is submitted in accordance with the regulations for the degree of Doctor of Philoso­ phy in the University of Glasgow. It is the record of research carried out at Glasgow University between October 1995 and September 1998. No part of it has previously been submitted by me for a degree at any university. I should like to express my gratitude to my supervisor, Prof. R.W.K.Odoni, for his ad­ vice and encouragement throughout the period of research. I would also like to thank Dr. S.D.Gohen for helping out with my supervision, and the E.P.S.R.G.
    [Show full text]
  • University of Groningen Panel Data Models Extended to Spatial Error
    University of Groningen Panel data models extended to spatial error autocorrelation or a spatially lagged dependent variable Elhorst, J. Paul IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document version below. Document Version Publisher's PDF, also known as Version of record Publication date: 2001 Link to publication in University of Groningen/UMCG research database Citation for published version (APA): Elhorst, J. P. (2001). Panel data models extended to spatial error autocorrelation or a spatially lagged dependent variable. s.n. Copyright Other than for strictly personal use, it is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license (like Creative Commons). The publication may also be distributed here under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license. More information can be found on the University of Groningen website: https://www.rug.nl/library/open-access/self-archiving-pure/taverne- amendment. Take-down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. Downloaded from the University of Groningen/UMCG research database (Pure): http://www.rug.nl/research/portal. For technical reasons the number of authors shown on this cover page is limited to 10 maximum. Download date: 01-10-2021 3$1(/'$7$02'(/6(;7(1'('7263$7,$/ (5525$872&255(/$7,2125$63$7,$//< /$**(''(3(1'(179$5,$%/( J.
    [Show full text]
  • Random Variable
    Random Variable Definition A random variable X on a sample spaceΩ is a real-valued function onΩ; that is, X :Ω !R.A discrete random variable is a random variable that takes on only a finite or countably infinite number of values. Discrete random variable X and real value a: the event \X = a" represents the set fs 2 Ω: X (s) = ag. X Pr(X = a) = Pr(s) s2Ω:X (s)=a Independence Definition Two random variables X and Y are independent if and only if Pr((X = x) \ (Y = y)) = Pr(X = x) · Pr(Y = y) for all values x and y. Similarly, random variables X1; X2;::: Xk are mutually independent if and only if for any subset I ⊆ [1; k] and any values xi ,i 2 I , ! \ Y Pr Xi = xi = Pr(Xi = xi ): i2I i2I Expectation Definition The expectation of a discrete random variable X , denoted by E[X ], is given by X E[X ] = i Pr(X = i); i where the summation is over all values in the range of X . The P expectation is finite if i jij Pr(X = i) converges; otherwise, the expectation is unbounded. The expectation (or mean or average) is a weighted sum over all possible values of the random variable. Median Definition The median of a random variable X is a value m such Pr(X < m) ≤ 1=2 and Pr(X > m) < 1=2: Linearity of Expectation Theorem For any two random variablesX andY E[X + Y ] = E[X ] + E[Y ]: Lemma For any constantc and discrete random variableX, E[cX ] = cE[X ]: Bernoulli Random Variable A Bernoulli or an indicator random variable: 1 if the experiment succeeds, Y = 0 otherwise.
    [Show full text]
  • Iterated Random Functions∗
    SIAM REVIEW c 1999 Society for Industrial and Applied Mathematics Vol. 41, No. 1, pp. 45–76 Iterated Random Functions∗ Persi Diaconisy David Freedmanz Abstract. Iterated random functions are used to draw pictures or simulate large Ising models, among other applications. They offer a method for studying the steady state distribution of a Markov chain, and give useful bounds on rates of convergence in a variety of examples. The present paper surveys the field and presents some new examples. There is a simple unifying idea: the iterates of random Lipschitz functions converge if the functions are contracting on the average. Key words. Markov chains, products of random matrices, iterated function systems, coupling from the past AMS subject classifications. 60J05, 60F05 PII. S0036144598338446 1. Introduction. The applied probability literature is nowadays quite daunting. Even relatively simple topics, like Markov chains, have generated enormous complex- ity. This paper describes a simple idea that helps to unify many arguments in Markov chains, simulation algorithms, control theory, queuing, and other branches of applied probability. The idea is that Markov chains can be constructed by iterating random functions on the state space S. More specifically, there is a family fθ : θ Θ of functions that map S into itself, and a probability distribution µ onf Θ. If the2 chaing is at x S, it moves by choosing θ at random from µ, and going to fθ(x). For now, µ does2 not depend on x. The process can be written as X0 = x0;X1 = fθ1 (x0);X2 =(fθ2 fθ1 )(x0); :::, with for composition of functions.
    [Show full text]
  • Automorphisms of Random Trees
    Automorphisms of Random Trees A Thesis Submitted to the Faculty of Drexel University by Le Yu in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Mathematics December 2012 c Copyright 2012 Le Yu. All Rights Reserved. ii Table of Contents LIST OF TABLES .................................................................... iii LIST OF FIGURES ................................................................... iv Abstract................................................................................ v 1. INTRODUCTION ................................................................. 1 2. LOG jAUT(T )j. ................................................................... 4 2.1 Product Representation ..................................................... 4 2.2 Excluding Large Branches .................................................. 10 2.3 Expected value of log jAut(T )j ............................................. 14 2.3.1 Expected value of N(T; d; B) ....................................... 16 2.3.2 Expected Value of log jAut(T )j .................................... 18 2.3.3 The unrooted case: En(log jAut(Un)j)............................. 23 2.4 Concentration around the mean ............................................ 26 2.4.1 Joyal's bijection...................................................... 27 2.4.2 Probability Spaces ................................................... 29 2.4.3 Application of the Azuma-McDiarmid theorem ................... 31 2.5 Evaluation of µ .............................................................
    [Show full text]
  • 1 Introduction 2 Gaussian Random Elements
    Gaussian Process in Statistical Modeling STA 941. Surya Tokdar 1 Introduction Gaussian distributions are important building blocks in statistical modeling. As mod- eling tools, they arise naturally from first principles reasonings based off the central limit theorem. They also provide the most natural way to incorporate dependency between multiple quantities. When there are infinitely many such quantities that can be viewed as a map or a function from an input space to an output space, the notion of Gaussianity continues to provide useful tools for constructing joint probability mod- els. In particular, there is a notion of a Gaussian probability distribution on infinite dimensional spaces such as RS, the space of all functions f : S ! R, where S is a subset of an Euclidean space. Below we build up a general framework of the notion of Gaussianity on arbitrary spaces. 2 Gaussian random elements As a starting point we only assume the concept of a univariate normality. A random variable X 2 R is said to be normally distributedp if there exist µ 2 R, σ > 0 such that the pdf of X is f(x) = exp{−(x − µ)2=(2σ2)g= 2πσ2, x 2 R. The distribution of X is N(µ, σ2). For completion, we allow the notation N(0; 0) to mean the degenerate distribution under which X = 0 with probability 1. 2.1 Formal definition Below we define the concept of a Gaussian random element in a Banach space F, a (complete) normed linear space with a norm k · k. Two specific examples of interest to us are •F = Rp for some p ≥ 1 equipped with the Euclidean norm •F = C(S), the space of all continuous functions from a compact S ⊂ Rd to R, equipped with the supremum norm kfk1 = sups2S jf(s)j.
    [Show full text]