Random Fields, Fall 2014 1
Total Page:16
File Type:pdf, Size:1020Kb
Random fields, Fall 2014 1 fMRI brain scan Ocean waves Nobel prize 2996 to The oceans cover 72% of John C. Mather the earth’s surface. George F. Smoot Essential for life on earth, “for discovery of the and huge economic blackbody form and importance through anisotropy of the cosmic fishing, transportation, microwave background oil and gas extraction radiation" PET brain scan TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A The course • Kolmogorov existence theorem, separable processes, measurable processes • Stationarity and isotropy • Orthogonal and spectral representations • Geometry • Exceedance sets • Rice formula • Slepian models Literature An unfinished manuscript “Applications of RANDOM FIELDS AND GEOMETRY: Foundations and Case Studies” by Robert Adler, Jonathan Taylor, and Keith Worsley. Complementary literature: “Level sets and extrema of random processes and fields” by Jean- Marc Azais and Mario Wschebor, Wiley, 2009 “Asymptotic Methods in the Theory of Gaussian Processes” by Vladimir Piterbarg, American Mathematical Society, ser. Translations of Mathematical Monographs, Vol. 148, 1995 “Random fields and Geometry” by Robert Adler and Jonathan Taylor, Springer 2007 Slides for ATW ch 2, p 24-39 Exercises: 2.8.1, 2.8.2, 2.8.3, 2.8.4, 2.8.5, 2.8.6 + excercises in slides Stochastic convergence (assumed known) 푎.푠. Almost sure convergence: 푋푛 X 퐿2 Mean square convergence: 푋푛 X 푃 Convergence in probability: 푋푛 X 푑 푑 푑 Convergence in distribution: 푋푛 X, 퐹푛 F, 푃푛 P, … 푎.푠. 푃 • 푋푛 X ⇒ 푋푛 푋 퐿2 푃 • 푋푛 X ⇒ 푋푛 푋 푃 퐿2 • 푋푛 푋 plus uniform integrability ⇒ 푋푛 X 푃 푎.푠. • 푋푛 푋 ⇒ there is a subsequence {푛푘} with 푋푛푘 X 푑 • The random variables don’t really mean anything for In particular, the 푋푛 and 푋 don’t need to be defined on the same probability space, and don’t need to have a simultaneous distribution Random field 푇 parameter space. In this course 푇 is 푹푁 some 푁 ≥ 1 or a subset (e.g. box or sphere or surface of sphere) of 푹푁 푹푑 value space An (푁, 푑) random field is a collection (or family) of random variables 푓푡; 푡 ∈ 푇 where 푇 is a set of dimension 푁 and the 푓푡 (or 푓(푡)) take values in 푹푑 푇 Or, a random function with values in 푹푑 푇 Or, a probablility measure on 푹푑 A realisation (or sample function, or sample path, or sample field, or observation, or trajectory, or …) is the function 푑 푓푡 휔 : 푇 푹 푡 ft(휔) for 휔 fixed. Two examples below: Microscopy image of Thresholded Gaussian tablet coating field Terminology random variable stochastic variable random element random vector stochastic element stochastic vector random field random process stochastic field stochastic process Finite dimensional distributions The finite-dimensional distribution functions of an (푁, 푑) random field {푓푡} are defined as 퐹푡1,…,푡푛 풙1, … , 풙푛 = 푃(푓푡1 ≤ 풙1, … , 푓푡푛 ≤ 풙n) and the family of finite-dimensional distribution functions is the set 푑 {퐹푡1,…,푡푛 풙1, … , 풙푛 ; 푡1, … , 푡푛 ∈ 푇, 풙1, … , 풙푛 ∈ 푅 , 푛 ≥ 1} This family has the following obvious properties: Symmetry: it is not changed under a simultaneous permutation of 푡1, … , 푡푛 and 풙1, … , 풙푛 Consistency: 퐹푡1,…,푡푛 풙1, … , 풙푛−1, ∞ = 퐹푡1,…,푡푛−1 풙1, … , 풙푛−1 Example of symmetry: 퐹푡1,푡2 풙1, 풙2 = 퐹푡2,푡1 풙2, 풙1 Example of consistency: Marginal distributions may be obtained from bivariate distributions, 퐹푡 풙 = 퐹푡,푠 풙, ∞ Three sample paths of a 1,1 random field. 퐹2,5,8 푥1, 푥2, 푥3 is the probability to obtain a sample path which passes through all three vertical lines (in a more general theory one instead of finite-dimensional distributions uses probabilities of cylindersets, 푃(푓푡1 ∈ 푩1, … , 푓푡푛 ∈ 푩n) ) (Daniell-)Kolmogorov extension theorem To any symmetric and consistent family of finite-dimensional distributions 푑 {퐹푡1,…,푡푛 풙1, … , 풙푛 ; 푡1, … , 푡푛 ∈ 푇, 풙1, … , 풙푛 ∈ 푅 , 푛 ≥ 1} there exists a probability triple (Ω, , 푃) and an 푁, 푑 B random field {푓푡; 푡 ∈ 푇 } which has these finite-dimensional distributions In the proof one takes 푑 푇 푑 푇 Ω = 푹 , = (푹 ) , 푇 and 푃 as the measure on (푹푑) which is uniquely determined by the finite-dimensional distribution. Thus an element of Ω is a function 푓: 푇 푅푑 which maps a point 푡 ∈ 푇 to the value 푓(푡). The field is defined as {푓푡 휔 = 푓 푡 ; 푡 ∈ 푇} Limitations of Kolmogorov’s theorem Many interesting sets, such as the set 퐶 = {휔; 푓푡 휔 is a continous function of 푡} 푇 do not belong to = (푹푑) , and hence, in Kolmogorov’s construction , the probababilityB of such events is not defined. One important way around this problem is to make a direct construction of the field on some other probability space (Ω, , 푃) where the interesting sets belong to , say 퐶 ∈ , so that their probabilities, say 푃(퐶), is well defined. And then, more fields are obtained as functions of the already constructed field! Modifications A field 푔푡; 푡 ∈ 푇 is a modification of the field 푓푡; 푡 ∈ 푇 if 푃 푔푡 = 푓푡 = 1, ∀푡 ∈ 푇 It is obvious (!) that 푔푡 has the same finite dimensional distributions as 푓푡. B The other common way to circumvent the limitation is to construct, on Kolmogorov’s (Ω, , 푃) a modification of 푓푡 which has the desired properties, say continuity. Whether this is possible or not (of course) dependes on which finite-dimensional distributions one is interested in. E.g. if they correspond to a Browninan motion it is possible, if they correspond to a Poisson process, it isn’t. Doob’s separability A field 푓푡; 푡 ∈ 푇 is separable if there is a countable subset 푇 푆 ∈ 푇 and a null set Λ ∈ (푹푑) such that for every closed set 퐵 ∈ 푹푑 and open set 퐼 ∈ 푇 it holds that 푓푡 휔 ∈ 퐵, ∀푡 ∈ 푆 ∩ 퐼 ⇒ 휔 ∈ ΛB or 푓푡 휔 ∈ 퐵, ∀푡 ∈ 퐼 A separable modification of a field always exists (at least for 푁 = 푑 = 1? ), and it can be seen that e.g. if a continuous modification of a field exists, then the separable modification is continuous. Example of modification: Ω = 0,1 , is the Borel sets on [0, 1], 푃 is Lebesgue measure, 푓푡 휔 = 0, ∀푡, 휔 and 0 if 푡 ≠ 휔 푔 휔 = 푡 1 if 푡 = 휔 Measurable fields A field 푓푡; 푡 ∈ 푇 is measurable if for almost all 휔 the sample path (function) 푑 푓 . 휔 : 푇 푅 B 푡 푓푡(휔) is (푅푑)-measurable (holds e.g. if the field is a.s. continuous). It then follows that the function of two variables 푓푡(휔) is measurable with respect to the product sigma-algebra × (푅푑), and one can then define integrals like 푇 ℎ 푓푡 푑푡 and use Fubini’s theorem for calculations like 퐸 ℎ 푓푡 푑푡 = 퐸(ℎ 푓푡 )푑푡 푇 푇 (above we have assumed that and (푅푑) are complete) ATW basically say that it is nice if one has seen the concepts of Kolmogorov extension, modification, and Doob separability, but that this has been taken care of once and for all by Kolmogorov, Doob and others, and that we shouldn’t worry about it any more in this course. And this is right (I hope). However, things are different for the theory of ”Empirical Processes”, the so far most efficient and high-tech tool to find asymptotic distributions of statistical estimators. In this theory, such ”measureability problems” pose important techical problems, and has formed much of the entire theory. Empirical process theory is closely related to the metods used to prove continuity and differentiability in this course. Gaussian fields (ATW p. 25-28) A random vector 퐗 = 푋1, … , 푋푑 has a multivariate Gaussian distribution iff one of the following conditions hold: 푑 • 훼, 푥 ≜ 푖 훼푖푋푖 has a univariate normal distribution for all 훼 ∈ 푅푑. • There exist a vector 풎 ∈ 푹푑 and a non-negative definite matrix 퐶 such that for all 휽 ∈ 푹푑 1 휙 휽 = 퐸 푒푖휽푋 = e푖휃풎−2휽퐶휽´ If 퐶 is positive definite and 푋 has the probability density 1 1 푒−2 풙−풎 퐶 풙−풎 ´ 2휋 푑 퐶 1/2 then 푋 is Gaussian. 풅 Here 푚 = 퐸 푿 and 퐶 = 퐶표푣 푿 . Similarly if the 푿풊 ∈ 푹 We write 퐗~푁푑 풎, 퐶 if 퐗 has a d-variate Gaussian distribution with mean 푚 and covariance matrix 퐶. Excercises (the first is (2.2.5), the second Exercise 2.8.2): (i) if 퐗~푁푑 풎, 퐶 and 퐴 is a 푑 × 푑 matrix, then 퐗퐴~푁푑 풎퐴, 퐴´퐶퐴 1 2 1 2 (ii) If 푿 = 푿 , 푿 with 푿 = 푋1, … , 푋푛 , 푿 = 푋푛+1, … , 푋푑 , with mean vectors 푚1 and 푚2 and covariance matrix 퐶1,1 퐶1,2 1 퐶 = , then the conditional distribution of 푿 given 퐶2,1 퐶2,2 2 푿 is n-variate normal with mean 1 2 2 −1 풎1|2 = 풎 + (푿 −풎 )퐶2,2 퐶2,1 and covariance matrix −1 퐶1|2 = 퐶1,1 − 퐶1,2퐶2,2 퐶2,1 A Gaussian random field is hence, by the Kolmogorov theorem, determined by its means and covariances Conversely, it also follows from the Kolmogorov theorem that given a function 푚: 푇 푹 and a non-negative definite function 퐶: 푇 × 푇 푹 there exist an 푁, 1 Gaussian random field which has 풎 as mean function and 퐶 as covariance function. 푁, 푑 Gaussian random fields for 푑 > 1 are the same, one just has to use more general notation. Gaussian related fields (ATW p. 28-30) An (푁, 푑) Gaussian related field 푓(푡); 푡 ∈ 푇 is defined from a (푁, 푘) Gaussian field 푔(푡); 푡 ∈ 푇 using a function 푘 푑 퐹: 푅 푅 by the formula 푓 푡 = 퐹 푔 푡 = 퐹 g1 t , … , 푔푘 푡 .