Probability, Random Processes, and Ergodic Properties

Total Page:16

File Type:pdf, Size:1020Kb

Probability, Random Processes, and Ergodic Properties Probability, Random Processes, and Ergodic Properties January 2, 2010 ii Probability, Random Processes, and Ergodic Properties Robert M. Gray Information Systems Laboratory Electrical Engineering Department Stanford University Springer-Verlag New York iv c 1987 by Springer Verlag. Revised 2001, 2006, 2007, 2008 by Robert M. Gray. v This book is affectionately dedicated to the memory of Elizabeth Dubois Jordan Gray 1906{1998 R. Adm. Augustine Heard Gray, U.S.N. 1888{1981 Sara Jean Dubois 1811{? and William \Old Billy" Gray 1750{1825 vi Contents Contents vii Preface ix 1 Probability and Random Processes 1 1.1 Introduction . .1 1.2 Probability Spaces and Random Variables . .1 1.3 Random Processes and Dynamical Systems . .6 1.4 Distributions . .8 1.5 Extension . 13 1.6 Isomorphism . 19 2 Standard alphabets 21 2.1 Extension of Probability Measures . 21 2.2 Standard Spaces . 22 2.3 Some properties of standard spaces . 26 2.4 Simple standard spaces . 29 2.5 Metric Spaces . 31 2.6 Extension in Standard Spaces . 36 2.7 The Kolmogorov Extension Theorem . 37 2.8 Extension Without a Basis . 38 3 Borel Spaces and Polish alphabets 45 3.1 Borel Spaces . 45 3.2 Polish Spaces . 48 3.3 Polish Schemes . 54 4 Averages 61 4.1 Introduction . 61 4.2 Discrete Measurements . 61 4.3 Quantization . 64 4.4 Expectation . 67 4.5 Time Averages . 77 4.6 Convergence of Random Variables . 80 4.7 Stationary Averages . 87 vii viii CONTENTS 5 Conditional Probability and Expectation 91 5.1 Introduction . 91 5.2 Measurements and Events . 91 5.3 Restrictions of Measures . 95 5.4 Elementary Conditional Probability . 95 5.5 Projections . 98 5.6 The Radon-Nikodym Theorem . 101 5.7 Conditional Probability . 104 5.8 Regular Conditional Probability . 106 5.9 Conditional Expectation . 109 5.10 Independence and Markov Chains . 115 6 Ergodic Properties 119 6.1 Ergodic Properties of Dynamical Systems . 119 6.2 Some Implications of Ergodic Properties . 122 6.3 Asymptotically Mean Stationary Processes . 127 6.4 Recurrence . 134 6.5 Asymptotic Mean Expectations . 138 6.6 Limiting Sample Averages . 140 6.7 Ergodicity . 142 7 Ergodic Theorems 149 7.1 Introduction . 149 7.2 The Pointwise Ergodic Theorem . 149 7.3 Block AMS Processes . 154 7.4 The Ergodic Decomposition . 156 7.5 The Subadditive Ergodic Theorem . 160 8 Process Metrics and the Ergodic Decomposition 169 8.1 Introduction . 169 8.2 A Metric Space of Measures . 170 8.3 The Rho-Bar Distance . 176 8.4 Measures on Measures . 182 8.5 The Ergodic Decomposition Revisited . 183 8.6 The Ergodic Decomposition of Markov Processes . 186 8.7 Barycenters . 188 8.8 Affine Functions of Measures . 191 8.9 The Ergodic Decomposition of Affine Functionals . 194 Bibliography 197 Index 201 Preface History and Goals This book has been written for several reasons, not all of which are academic. This material was for many years the first half of a book in progress on information and ergodic theory. The intent was and is to provide a reasonably self-contained advanced treatment of measure theory, probability theory, and the theory of discrete time random processes with an emphasis on general alphabets and on ergodic and stationary properties of random processes that might be neither ergodic nor stationary. The intended audience was mathematically inclined engineering graduate students and visiting scholars who had not had formal courses in measure theoretic probability. Much of the material is familiar stuff for mathematicians, but many of the topics and results have not previously appeared in books. The original project grew too large and the first part contained much that would likely bore mathematicians and discourage them from the second part. Hence I finally followed a suggestion to separate the material and split the project in two. The original justification for the present manuscript was the pragmatic one that it would be a shame to waste all the effort thus far expended. A more idealistic motivation was that the presentation had merit as filling a unique, albeit small, hole in the literature. Personal experience indicates that the intended audience rarely has the time to take a complete course in measure and probability theory in a mathematics or statistics department, at least not before they need some of the material in their research. In addition, many of the existing mathematical texts on the subject are hard for this audience to follow, and the emphasis is not well matched to engineering applications. A notable exception is Ash's excellent text [1], which was likely influenced by his original training as an electrical engineer. Still, even that text devotes little effort to ergodic theorems, perhaps the most fundamentally important family of results for applying probability theory to real problems. In addition, there are many other special topics that are given little space (or none at all) in most texts on advanced probability and random processes. Examples of topics developed in more depth here than in most existing texts are the following: Random processes with standard alphabets We develop the theory of standard spaces as a model of quite general process alphabets. Although not as general (or abstract) as examples often considered by probability theorists, standard spaces have useful structural properties that simplify the proofs of some general results and yield additional results that may not hold in the more general abstract case. Three important examples of results holding for standard alphabets that have not been proved in the general abstract case are the Kolmogorov extension theorem, the ergodic decomposition, and the existence of regular conditional probabilities. In fact, Blackwell [6] introduced the notion of a Lusin space, a structure closely related to a standard space, in order to avoid known examples of probability spaces where the Kolmogorov extension theorem does not hold and regular conditional probabilities do not exist. Standard ix x PREFACE spaces include the common models of finite alphabets (digital processes) and real alphabets as well as more general complete separable metric spaces (Polish spaces). Thus they include many function spaces, Euclidean vector spaces, two-dimensional image intensity rasters, etc. The basic theory of standard Borel spaces may be found in the elegant text of Parthasarathy [55], and treatments of standard spaces and the related Lusin and Suslin spaces may be found in Christensen [10], Schwartz [62], Bourbaki [7], and Cohn [12]. We here provide a different and more coding oriented development of the basic results and attempt to separate clearly the properties of standard spaces, which are useful and easy to manipulate, from the demonstra- tions that certain spaces are standard, which are more complicated and can be skipped. Thus, unlike in the traditional treatments, we define and study standard spaces first from a purely probability theory point of view and postpone the topological metric space considerations until later. Nonstationary and nonergodic processes We develop the theory of asymptotically mean sta- tionary processes and the ergodic decomposition in order to model many physical processes better than can traditional stationary and ergodic processes. Both topics are virtually absent in all books on random processes, yet they are fundamental to understanding the limiting behavior of nonergodic and nonstationary processes. Both topics are considered in Krengel's excellent book on ergodic theorems [41], but the treatment here is more detailed and in greater depth. We consider both the common two-sided processes, which are considered to have been producing outputs forever, and the more difficult one-sided processes, which better model processes that are \turned on" at some specific time and which exhibit transient behavior. Ergodic properties and theorems We develop the notion of time averages along with that of probabilistic averages to emphasize their similarity and to demonstrate many of the implica- tions of the existence of limiting sample averages. We prove the ergodic theorem theorem for the general case of asymptotically mean stationary processes. In fact, it is shown that asymp- totic mean stationarity is both sufficient and necessary for the classical pointwise or almost everywhere ergodic theorem to hold for all bounded measurements. We also prove the sub- additive ergodic theorem of Kingman [39], which is useful for studying the limiting behavior of certain measurements on random processes that are not simple arithmetic averages. The proofs are based on recent simple proofs of the ergodic theorem developed by Ornstein and Weiss [52], Katznelson and Weiss [38], Jones [37], and Shields [64]. These proofs use coding arguments reminiscent of information and communication theory rather than the traditional (and somewhat tricky) maximal ergodic theorem. We consider the interrelations of stationary and ergodic properties of processes that are stationary or ergodic with respect to block shifts, that is, processes that produce stationary or ergodic vectors rather than scalars | a topic largely developed by Nedoma [49] which plays an important role in the general versions of Shannon channel and source coding theorems. Process distance measures We develop measures of a \distance" between random processes. Such results quantify how \close" one process is to another and are useful for considering spaces of random processes. These in turn provide the means of proving the ergodic decomposition of certain functionals of random processes and of characterizing how close or different the long term behavior of distinct random processes can be expected to be. Of particular interest are the distribution or variational distance and the Kantorovich/Vasershtein/Ornstein matching distance. Having described the topics treated here that are lacking in most texts, we admit to the omission of many topics usually contained in advanced texts on random processes or second books on random PREFACE xi processes for engineers. The most obvious omission is that of continuous time random processes.
Recommended publications
  • Measure-Theoretic Probability I
    Measure-Theoretic Probability I Steven P.Lalley Winter 2017 1 1 Measure Theory 1.1 Why Measure Theory? There are two different views – not necessarily exclusive – on what “probability” means: the subjectivist view and the frequentist view. To the subjectivist, probability is a system of laws that should govern a rational person’s behavior in situations where a bet must be placed (not necessarily just in a casino, but in situations where a decision must be made about how to proceed when only imperfect information about the outcome of the decision is available, for instance, should I allow Dr. Scissorhands to replace my arthritic knee by a plastic joint?). To the frequentist, the laws of probability describe the long- run relative frequencies of different events in “experiments” that can be repeated under roughly identical conditions, for instance, rolling a pair of dice. For the frequentist inter- pretation, it is imperative that probability spaces be large enough to allow a description of an experiment, like dice-rolling, that is repeated infinitely many times, and that the mathematical laws should permit easy handling of limits, so that one can make sense of things like “the probability that the long-run fraction of dice rolls where the two dice sum to 7 is 1/6”. But even for the subjectivist, the laws of probability should allow for description of situations where there might be a continuum of possible outcomes, or pos- sible actions to be taken. Once one is reconciled to the need for such flexibility, it soon becomes apparent that measure theory (the theory of countably additive, as opposed to merely finitely additive measures) is the only way to go.
    [Show full text]
  • Stationary Processes and Their Statistical Properties
    Stationary Processes and Their Statistical Properties Brian Borchers March 29, 2001 1 Stationary processes A discrete time stochastic process is a sequence of random variables Z1, Z2, :::. In practice we will typically analyze a single realization z1, z2, :::, zn of the stochastic process and attempt to esimate the statistical properties of the stochastic process from the realization. We will also consider the problem of predicting zn+1 from the previous elements of the sequence. We will begin by focusing on the very important class of stationary stochas- tic processes. A stochastic process is strictly stationary if its statistical prop- erties are unaffected by shifting the stochastic process in time. In particular, this means that if we take a subsequence Zk+1, :::, Zk+m, then the joint distribution of the m random variables will be the same no matter what k is. Stationarity requires that the mean of the stochastic process be a constant. E[Zk] = µ. and that the variance is constant 2 V ar[Zk] = σZ : Also, stationarity requires that the covariance of two elements separated by a distance m is constant. That is, Cov(Zk;Zk+m) is constant. This covariance is called the autocovariance at lag m, and we will use the notation γm. Since Cov(Zk;Zk+m) = Cov(Zk+m;Zk), we need only find γm for m 0. The ≥ correlation of Zk and Zk+m is the autocorrelation at lag m. We will use the notation ρm for the autocorrelation. It is easy to show that γk ρk = : γ0 1 2 The autocovariance and autocorrelation ma- trices The covariance matrix for the random variables Z1, :::, Zn is called an auto- covariance matrix.
    [Show full text]
  • Methods of Monte Carlo Simulation II
    Methods of Monte Carlo Simulation II Ulm University Institute of Stochastics Lecture Notes Dr. Tim Brereton Summer Term 2014 Ulm, 2014 2 Contents 1 SomeSimpleStochasticProcesses 7 1.1 StochasticProcesses . 7 1.2 RandomWalks .......................... 7 1.2.1 BernoulliProcesses . 7 1.2.2 RandomWalks ...................... 10 1.2.3 ProbabilitiesofRandomWalks . 13 1.2.4 Distribution of Xn .................... 13 1.2.5 FirstPassageTime . 14 2 Estimators 17 2.1 Bias, Variance, the Central Limit Theorem and Mean Square Error................................ 19 2.2 Non-AsymptoticErrorBounds. 22 2.3 Big O and Little o Notation ................... 23 3 Markov Chains 25 3.1 SimulatingMarkovChains . 28 3.1.1 Drawing from a Discrete Uniform Distribution . 28 3.1.2 Drawing From A Discrete Distribution on a Small State Space ........................... 28 3.1.3 SimulatingaMarkovChain . 28 3.2 Communication .......................... 29 3.3 TheStrongMarkovProperty . 30 3.4 RecurrenceandTransience . 31 3.4.1 RecurrenceofRandomWalks . 33 3.5 InvariantDistributions . 34 3.6 LimitingDistribution. 36 3.7 Reversibility............................ 37 4 The Poisson Process 39 4.1 Point Processes on [0, )..................... 39 ∞ 3 4 CONTENTS 4.2 PoissonProcess .......................... 41 4.2.1 Order Statistics and the Distribution of Arrival Times 44 4.2.2 DistributionofArrivalTimes . 45 4.3 SimulatingPoissonProcesses. 46 4.3.1 Using the Infinitesimal Definition to Simulate Approx- imately .......................... 46 4.3.2 SimulatingtheArrivalTimes . 47 4.3.3 SimulatingtheInter-ArrivalTimes . 48 4.4 InhomogenousPoissonProcesses. 48 4.5 Simulating an Inhomogenous Poisson Process . 49 4.5.1 Acceptance-Rejection. 49 4.5.2 Infinitesimal Approach (Approximate) . 50 4.6 CompoundPoissonProcesses . 51 5 ContinuousTimeMarkovChains 53 5.1 TransitionFunction. 53 5.2 InfinitesimalGenerator . 54 5.3 ContinuousTimeMarkovChains .
    [Show full text]
  • Probability and Statistics Lecture Notes
    Probability and Statistics Lecture Notes Antonio Jiménez-Martínez Chapter 1 Probability spaces In this chapter we introduce the theoretical structures that will allow us to assign proba- bilities in a wide range of probability problems. 1.1. Examples of random phenomena Science attempts to formulate general laws on the basis of observation and experiment. The simplest and most used scheme of such laws is: if a set of conditions B is satisfied =) event A occurs. Examples of such laws are the law of gravity, the law of conservation of mass, and many other instances in chemistry, physics, biology... If event A occurs inevitably whenever the set of conditions B is satisfied, we say that A is certain or sure (under the set of conditions B). If A can never occur whenever B is satisfied, we say that A is impossible (under the set of conditions B). If A may or may not occur whenever B is satisfied, then A is said to be a random phenomenon. Random phenomena is our subject matter. Unlike certain and impossible events, the presence of randomness implies that the set of conditions B do not reflect all the necessary and sufficient conditions for the event A to occur. It might seem them impossible to make any worthwhile statements about random phenomena. However, experience has shown that many random phenomena exhibit a statistical regularity that makes them subject to study. For such random phenomena it is possible to estimate the chance of occurrence of the random event. This estimate can be obtained from laws, called probabilistic or stochastic, with the form: if a set of conditions B is satisfied event A occurs m times =) repeatedly n times out of the n repetitions.
    [Show full text]
  • Ergodicity, Decisions, and Partial Information
    Ergodicity, Decisions, and Partial Information Ramon van Handel Abstract In the simplest sequential decision problem for an ergodic stochastic pro- cess X, at each time n a decision un is made as a function of past observations X0,...,Xn 1, and a loss l(un,Xn) is incurred. In this setting, it is known that one may choose− (under a mild integrability assumption) a decision strategy whose path- wise time-average loss is asymptotically smaller than that of any other strategy. The corresponding problem in the case of partial information proves to be much more delicate, however: if the process X is not observable, but decisions must be based on the observation of a different process Y, the existence of pathwise optimal strategies is not guaranteed. The aim of this paper is to exhibit connections between pathwise optimal strategies and notions from ergodic theory. The sequential decision problem is developed in the general setting of an ergodic dynamical system (Ω,B,P,T) with partial information Y B. The existence of pathwise optimal strategies grounded in ⊆ two basic properties: the conditional ergodic theory of the dynamical system, and the complexity of the loss function. When the loss function is not too complex, a gen- eral sufficient condition for the existence of pathwise optimal strategies is that the dynamical system is a conditional K-automorphism relative to the past observations n n 0 T Y. If the conditional ergodicity assumption is strengthened, the complexity assumption≥ can be weakened. Several examples demonstrate the interplay between complexity and ergodicity, which does not arise in the case of full information.
    [Show full text]
  • On the Continuing Relevance of Mandelbrot's Non-Ergodic Fractional Renewal Models of 1963 to 1967
    Nicholas W. Watkins On the continuing relevance of Mandelbrot's non-ergodic fractional renewal models of 1963 to 1967 Article (Published version) (Refereed) Original citation: Watkins, Nicholas W. (2017) On the continuing relevance of Mandelbrot's non-ergodic fractional renewal models of 1963 to 1967.European Physical Journal B, 90 (241). ISSN 1434-6028 DOI: 10.1140/epjb/e2017-80357-3 Reuse of this item is permitted through licensing under the Creative Commons: © 2017 The Author CC BY 4.0 This version available at: http://eprints.lse.ac.uk/84858/ Available in LSE Research Online: February 2018 LSE has developed LSE Research Online so that users may access research output of the School. Copyright © and Moral Rights for the papers on this site are retained by the individual authors and/or other copyright owners. You may freely distribute the URL (http://eprints.lse.ac.uk) of the LSE Research Online website. Eur. Phys. J. B (2017) 90: 241 DOI: 10.1140/epjb/e2017-80357-3 THE EUROPEAN PHYSICAL JOURNAL B Regular Article On the continuing relevance of Mandelbrot's non-ergodic fractional renewal models of 1963 to 1967? Nicholas W. Watkins1,2,3 ,a 1 Centre for the Analysis of Time Series, London School of Economics and Political Science, London, UK 2 Centre for Fusion, Space and Astrophysics, University of Warwick, Coventry, UK 3 Faculty of Science, Technology, Engineering and Mathematics, Open University, Milton Keynes, UK Received 20 June 2017 / Received in final form 25 September 2017 Published online 11 December 2017 c The Author(s) 2017. This article is published with open access at Springerlink.com Abstract.
    [Show full text]
  • Examples of Stationary Time Series
    Statistics 910, #2 1 Examples of Stationary Time Series Overview 1. Stationarity 2. Linear processes 3. Cyclic models 4. Nonlinear models Stationarity Strict stationarity (Defn 1.6) Probability distribution of the stochastic process fXtgis invariant under a shift in time, P (Xt1 ≤ x1;Xt2 ≤ x2;:::;Xtk ≤ xk) = F (xt1 ; xt2 ; : : : ; xtk ) = F (xh+t1 ; xh+t2 ; : : : ; xh+tk ) = P (Xh+t1 ≤ x1;Xh+t2 ≤ x2;:::;Xh+tk ≤ xk) for any time shift h and xj. Weak stationarity (Defn 1.7) (aka, second-order stationarity) The mean and autocovariance of the stochastic process are finite and invariant under a shift in time, E Xt = µt = µ Cov(Xt;Xs) = E (Xt−µt)(Xs−µs) = γ(t; s) = γ(t−s) The separation rather than location in time matters. Equivalence If the process is Gaussian with finite second moments, then weak stationarity is equivalent to strong stationarity. Strict stationar- ity implies weak stationarity only if the necessary moments exist. Relevance Stationarity matters because it provides a framework in which averaging makes sense. Unless properties like the mean and covariance are either fixed or \evolve" in a known manner, we cannot average the observed data. What operations produce a stationary process? Can we recognize/identify these in data? Statistics 910, #2 2 Moving Average White noise Sequence of uncorrelated random variables with finite vari- ance, ( 2 often often σw = 1 if t = s; E Wt = µ = 0 Cov(Wt;Ws) = 0 otherwise The input component (fXtg in what follows) is often modeled as white noise. Strict white noise replaces uncorrelated by independent. Moving average A stochastic process formed by taking a weighted average of another time series, often formed from white noise.
    [Show full text]
  • Ergodic Theory
    MATH41112/61112 Ergodic Theory Charles Walkden 4th January, 2018 MATH4/61112 Contents Contents 0 Preliminaries 2 1 An introduction to ergodic theory. Uniform distribution of real se- quences 4 2 More on uniform distribution mod 1. Measure spaces. 13 3 Lebesgue integration. Invariant measures 23 4 More examples of invariant measures 38 5 Ergodic measures: definition, criteria, and basic examples 43 6 Ergodic measures: Using the Hahn-Kolmogorov Extension Theorem to prove ergodicity 53 7 Continuous transformations on compact metric spaces 62 8 Ergodic measures for continuous transformations 72 9 Recurrence 83 10 Birkhoff’s Ergodic Theorem 89 11 Applications of Birkhoff’s Ergodic Theorem 99 12 Solutions to the Exercises 108 1 MATH4/61112 0. Preliminaries 0. Preliminaries 0.1 Contact details § The lecturer is Dr Charles Walkden, Room 2.241, Tel: 0161 275 5805, Email: [email protected]. My office hour is: Monday 2pm-3pm. If you want to see me at another time then please email me first to arrange a mutually convenient time. 0.2 Course structure § This is a reading course, supported by one lecture per week. I have split the notes into weekly sections. You are expected to have read through the material before the lecture, and then go over it again afterwards in your own time. In the lectures I will highlight the most important parts, explain the statements of the theorems and what they mean in practice, and point out common misunderstandings. As a general rule, I will not normally go through the proofs in great detail (but they are examinable unless indicated otherwise).
    [Show full text]
  • Stationary Processes
    Stationary processes Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania [email protected] http://www.seas.upenn.edu/users/~aribeiro/ November 25, 2019 Stoch. Systems Analysis Stationary processes 1 Stationary stochastic processes Stationary stochastic processes Autocorrelation function and wide sense stationary processes Fourier transforms Linear time invariant systems Power spectral density and linear filtering of stochastic processes Stoch. Systems Analysis Stationary processes 2 Stationary stochastic processes I All probabilities are invariant to time shits, i.e., for any s P[X (t1 + s) ≥ x1; X (t2 + s) ≥ x2;:::; X (tK + s) ≥ xK ] = P[X (t1) ≥ x1; X (t2) ≥ x2;:::; X (tK ) ≥ xK ] I If above relation is true process is called strictly stationary (SS) I First order stationary ) probs. of single variables are shift invariant P[X (t + s) ≥ x] = P [X (t) ≥ x] I Second order stationary ) joint probs. of pairs are shift invariant P[X (t1 + s) ≥ x1; X (t2 + s) ≥ x2] = P [X (t1) ≥ x1; X (t2) ≥ x2] Stoch. Systems Analysis Stationary processes 3 Pdfs and moments of stationary process I For SS process joint cdfs are shift invariant. Whereby, pdfs also are fX (t+s)(x) = fX (t)(x) = fX (0)(x) := fX (x) I As a consequence, the mean of a SS process is constant Z 1 Z 1 µ(t) := E [X (t)] = xfX (t)(x) = xfX (x) = µ −∞ −∞ I The variance of a SS process is also constant Z 1 Z 1 2 2 2 var [X (t)] := (x − µ) fX (t)(x) = (x − µ) fX (x) = σ −∞ −∞ I The power of a SS process (second moment) is also constant Z 1 Z 1 2 2 2 2 2 E X (t) := x fX (t)(x) = x fX (x) = σ + µ −∞ −∞ Stoch.
    [Show full text]
  • Ergodicity and Metric Transitivity
    Chapter 25 Ergodicity and Metric Transitivity Section 25.1 explains the ideas of ergodicity (roughly, there is only one invariant set of positive measure) and metric transivity (roughly, the system has a positive probability of going from any- where to anywhere), and why they are (almost) the same. Section 25.2 gives some examples of ergodic systems. Section 25.3 deduces some consequences of ergodicity, most im- portantly that time averages have deterministic limits ( 25.3.1), and an asymptotic approach to independence between even§ts at widely separated times ( 25.3.2), admittedly in a very weak sense. § 25.1 Metric Transitivity Definition 341 (Ergodic Systems, Processes, Measures and Transfor- mations) A dynamical system Ξ, , µ, T is ergodic, or an ergodic system or an ergodic process when µ(C) = 0 orXµ(C) = 1 for every T -invariant set C. µ is called a T -ergodic measure, and T is called a µ-ergodic transformation, or just an ergodic measure and ergodic transformation, respectively. Remark: Most authorities require a µ-ergodic transformation to also be measure-preserving for µ. But (Corollary 54) measure-preserving transforma- tions are necessarily stationary, and we want to minimize our stationarity as- sumptions. So what most books call “ergodic”, we have to qualify as “stationary and ergodic”. (Conversely, when other people talk about processes being “sta- tionary and ergodic”, they mean “stationary with only one ergodic component”; but of that, more later. Definition 342 (Metric Transitivity) A dynamical system is metrically tran- sitive, metrically indecomposable, or irreducible when, for any two sets A, B n ∈ , if µ(A), µ(B) > 0, there exists an n such that µ(T − A B) > 0.
    [Show full text]
  • MA427 Ergodic Theory
    MA427 Ergodic Theory Course Notes (2012-13) 1 Introduction 1.1 Orbits Let X be a mathematical space. For example, X could be the unit interval [0; 1], a circle, a torus, or something far more complicated like a Cantor set. Let T : X ! X be a function that maps X into itself. Let x 2 X be a point. We can repeatedly apply the map T to the point x to obtain the sequence: fx; T (x);T (T (x));T (T (T (x))); : : : ; :::g: We will often write T n(x) = T (··· (T (T (x)))) (n times). The sequence of points x; T (x);T 2(x);::: is called the orbit of x. We think of applying the map T as the passage of time. Thus we think of T (x) as where the point x has moved to after time 1, T 2(x) is where the point x has moved to after time 2, etc. Some points x 2 X return to where they started. That is, T n(x) = x for some n > 1. We say that such a point x is periodic with period n. By way of contrast, points may move move densely around the space X. (A sequence is said to be dense if (loosely speaking) it comes arbitrarily close to every point of X.) If we take two points x; y of X that start very close then their orbits will initially be close. However, it often happens that in the long term their orbits move apart and indeed become dramatically different.
    [Show full text]
  • Conformal Fractals – Ergodic Theory Methods
    Conformal Fractals – Ergodic Theory Methods Feliks Przytycki Mariusz Urba´nski May 17, 2009 2 Contents Introduction 7 0 Basic examples and definitions 15 1 Measure preserving endomorphisms 25 1.1 Measurespacesandmartingaletheorem . 25 1.2 Measure preserving endomorphisms, ergodicity . .. 28 1.3 Entropyofpartition ........................ 34 1.4 Entropyofendomorphism . 37 1.5 Shannon-Mcmillan-Breimantheorem . 41 1.6 Lebesguespaces............................ 44 1.7 Rohlinnaturalextension . 48 1.8 Generalized entropy, convergence theorems . .. 54 1.9 Countabletoonemaps.. .. .. .. .. .. .. .. .. .. 58 1.10 Mixingproperties . .. .. .. .. .. .. .. .. .. .. 61 1.11 Probabilitylawsand Bernoulliproperty . 63 Exercises .................................. 68 Bibliographicalnotes. 73 2 Compact metric spaces 75 2.1 Invariantmeasures . .. .. .. .. .. .. .. .. .. .. 75 2.2 Topological pressure and topological entropy . ... 83 2.3 Pressureoncompactmetricspaces . 87 2.4 VariationalPrinciple . 89 2.5 Equilibrium states and expansive maps . 94 2.6 Functionalanalysisapproach . 97 Exercises ..................................106 Bibliographicalnotes. 109 3 Distance expanding maps 111 3.1 Distance expanding open maps, basic properties . 112 3.2 Shadowingofpseudoorbits . 114 3.3 Spectral decomposition. Mixing properties . 116 3.4 H¨older continuous functions . 122 3 4 CONTENTS 3.5 Markov partitions and symbolic representation . 127 3.6 Expansive maps are expanding in some metric . 134 Exercises .................................. 136 Bibliographicalnotes. 138 4
    [Show full text]