Proof-Theoretic Strength and Indicator Arguments

Total Page:16

File Type:pdf, Size:1020Kb

Proof-Theoretic Strength and Indicator Arguments Preliminaries Indicators Proof-theoretic strength and indicator arguments Keita Yokoyama Japan Advanced Institute of Science and Technology September 16, 2016 Keita Yokoyama Proof-theoretic strength and indicator arguments 1 / 18 Preliminaries Indicators Indicators What is indicator? It is introduced by Kirby and Paris in 1970’s, and the general frame work is given by Kaye. It is used to prove the independence of the Paris-Harrington principle from PA. A tool to study cuts of nonstandard models of arithmetic. Indicators are useful to analyze the proof-theoretic strength of combinatorial statements in arithmetic. Note that most theorems in this talk are more or less folklores (in the field of nonstandard models of arithmetic). Keita Yokoyama Proof-theoretic strength and indicator arguments 2 / 18 Preliminaries Indicators Nonstandard models of arithmetic In this talk we will mainly use the base system EFA = I∆0 + exp or ∗ 0 0 RCA0, which consists of I∆0 + exp plus ∆1-comprehension, and models we will consider will be countable nonstandard. Let M j= EFA. I ⊆ M is said to be a cut (abbr. I ⊆e M) if a < b 2 I ! a 2 I and I is closed under addition + and multiplication ·. Cod(M) = fX ⊆ M j X is M-finiteg, where M-finite set is a set coded by an element in M (by means of the usual binary coding). for Z 2 Cod(M), jZj denotes the internal cardinality of Z in M. for I ⊆e M, Cod(M=I) := fX \ I j X 2 Cod(M)g. Proposition If I ⊆e M, then I is a Σ0-elementary substructure of M. Keita Yokoyama Proof-theoretic strength and indicator arguments 3 / 18 Preliminaries Indicators Cuts There are several important types of cuts. Theorem (exponentially closed cut, Simpson/Smith) Let M j= EFA, and let I (e M. Then the following are equivalent. ∗ 1 ; = j (I Cod(M I)) = WKL0. 2 I is closed under exp. Theorem (semi-regular cut) Let M j= EFA, and let I (e M. Then the following are equivalent. 1 (I; Cod(M=I)) j= WKL0. 2 I is semi-regular, i.e., if X 2 Cod(M) and jXj 2 I, then X \ I is bounded in I. These combinatorial characterization of cuts play key roles in the definition of indicators. Keita Yokoyama Proof-theoretic strength and indicator arguments 4 / 18 Preliminaries Indicators Indicators Generalization Indicators Let T be a theory of second-order arithmetic. 2 A Σ0-definable function Y :[M] ! M is said to be an indicator for ⊇ ∗ T WKL0 if Y(x; y) ≤ y, if x0 ≤ x < y ≤ y0, then Y(x; y) ≤ Y(x0; y0), Y(x; y) > ! if and only if there exists a cut I ⊆e M such that x 2 I < y and (I; Cod(M=I)) j= T. (Here, Y(x; y) > ! means that Y(x; y) > n for any standard natural number n.) Example ; f n ≤ g ∗ Y(x y) = max n : exp (x) y is an indicator for WKL0. Y(x; y) = maxfn :any f[[x; y]]n ! 2 has a homogeneous set Z ⊆ [x; y] such that jZj > min Zg is an indicator for ACA0. Keita Yokoyama Proof-theoretic strength and indicator arguments 5 / 18 Preliminaries Indicators Indicators Generalization Basic properties of indicators Theorem If Y is an indicator for a theory T, then for any n 2 !, T ` 8x9yY(x; y) ≥ n. Theorem 0 If Y is an indicator for a theory T, then, T is a Π2-conservative extension of EFA + f8x9yY(x; y) ≥ n j n 2 !g. Y f j ; ≥ g Let Fn (x) = min y Y(x y) n . Theorem If Y is an indicator for a theory T and T ` 8x9yθ(x; y) for some Σ1-formula θ, then, there exists n 2 ! such that ` 8 9 < Y θ ; T x y Fn (x) (x y). Keita Yokoyama Proof-theoretic strength and indicator arguments 6 / 18 Preliminaries Indicators Indicators Generalization Let Fk be the k-th fast-growing function. Example Y(x; y) = maxfk : Fk (x) < yg is an indicator for WKL0. Thus, we have WKL0 ` 8x9yFk (x) < y for any k 2 !, if WKL0 ` 8x9yθ(x; y) then there exists some k 2 ! such that WKL0 ` 8x9y < Fk (x)θ(x; y), the proof-theoretic strength of WKL0 is the same as the totality of all primitive recursive functions. Once you find an indicator for a theory T, one can characterize its 0 Π2-part. Theorem ⊇ ∗ Any consistent recursive theory T WKL0 (or first-order theory extending EFA) has an indicator. Keita Yokoyama Proof-theoretic strength and indicator arguments 7 / 18 Preliminaries Indicators Indicators Generalization Set indicators One can generalize indicators to capture wider class of formulas. Let T be a theory of second-order arithmetic. A Σ0-definable function Y : Cod(M) ! M is said to be a set ⊇ ∗ indicator for T WKL0 if Y(F) ≤ max F, if F ⊆ F0, then Y(F) ≤ Y(F0), Y(F) > ! if and only if there exists a cut I ⊆e M such that min F 2 I < max F and (I; Cod(M=I)) j= T, and F \ I is unbounded in I. Note that if Y is a set indicator, then Y 0(x; y) = Y([x; y]) is an indicator function. Example m Y(F) = maxfm : F is ! -largeg is an indicator for WKL0. Keita Yokoyama Proof-theoretic strength and indicator arguments 8 / 18 Preliminaries Indicators Indicators Generalization Basic properties of indicators (review) Theorem If Y is an indicator for a theory T, then for any n 2 !, T ` 8x9yY(x; y) ≥ n. Theorem 0 If Y is an indicator for a theory T, then, T is a Π2-conservative extension of EFA + f8x9yY(x; y) ≥ n j n 2 !g. Y f j ; ≥ g Let Fn (x) = min y Y(x y) n . Theorem If Y is an indicator for a theory T and T ` 8x9yθ(x; y) for some Σ1-formula θ, then, there exists n 2 ! such that ` 8 9 < Y θ ; T x y Fn (x) (x y). Keita Yokoyama Proof-theoretic strength and indicator arguments 9 / 18 Preliminaries Indicators Indicators Generalization Basic properties of set indicators Theorem If Y is a set indicator for a theory T, then for any n 2 !, T ` 8X ⊆inf N 9F ⊆fin X(Y(F) ≥ n). Theorem ~ 0 If Y is a set indicator for a theory T, then, T is a Π3-conservative ∗ f8 ⊆ N 9 ⊆ ≥ j 2 !g extension of RCA0 + X inf F fin X(Y(F) n) n . ~ 0 8 0 Here, a Π3-formula is of the form X (X) where is Π3. Theorem If Y is a set indicator for a theory T and T ` 8X ⊆inf N 9F ⊆fin Xθ(F) for some Σ1-formula θ, then, there exists n 2 ! such that T ` 8Z ⊆fin N(Y(Z) ≥ n ! 9F ⊆ Z θ(F)). Keita Yokoyama Proof-theoretic strength and indicator arguments 10 / 18 Preliminaries Indicators Indicators Generalization Example m Y(F) = maxfm : F is ! -largeg is an indicator for WKL0. Thus, ~ 0 all the Π3-consequences of WKL0 can be captured by !m-largeness notion. Question Is there a canonical way to find indicators? Keita Yokoyama Proof-theoretic strength and indicator arguments 11 / 18 Preliminaries Indicators Indicators Generalization Ramsey-like statements Definition (Ramsey-like formulas) 1 1 A Ramsey-like-Π2-formula is a Π2-formula of the form (8f :[N]n ! k)(9Y)(Y is infinite ^ Ψ(f; Y)) n where Ψ(f; Y) is of the form (8G ⊆fin Y)Ψ0(f ↾ [[0; max G]N] ; G) 0 such that Ψ0 is a ∆0-formula. n 1 In particular, RTk is a Ramsey-like-Π2-statement where Ψ(f; Y) is the formula “Y is homogeneous for f”. Theorem 1 8 9 ; Any restricted Π2-formula of the form X YΘ(X Y) where Θ is a 0 Σ3-formula is equivalent to a Ramsey-like formula over WKL0. Note that this theorem can be proved by a canonical syntactical calculation. Keita Yokoyama Proof-theoretic strength and indicator arguments 12 / 18 Preliminaries Indicators Indicators Generalization Density Definition (EFA, Density notion) Given a Ramsey-like formula Γ = (8f :[N]n ! k)(9Y)(Y is infinite ^ Ψ(f; Y)), Z ⊆fin N is said to be 0-dense(Γ) if jZj; min Z > 2, Z ⊆fin N is said to be (m + 1)-dense(Γ) if (for any n; k < min Z and) for any f : [[0; max Z]]n ! k, there is an m-dense(Γ) set Y ⊆ Z such that Ψ(f; Y) holds, and, for any partition Z0 t · · · t Z`−1 = Z such that ` ≤ Z0 < ··· < Z`−1, one of Zi’s is m-dense(Γ). Note that “Z is m-dense(Γ)” can be expressed by a ∆0-formula. Put YΓ(F) := maxfm j F is m-dense(Γ)g. Theorem YΓ is a set indicator for WKL0 + Γ. Keita Yokoyama Proof-theoretic strength and indicator arguments 13 / 18 Preliminaries Indicators Indicators Generalization Characterizing proof-theoretic strength by indicators One can characterize the proof-theoretic strength of a finite 1 ⊇ restricted Π2-theory T WKL0 as follows. Find a Ramsey-like formula Γ such that T $ WKL0 + Γ. ~ 0 Then, m-dense(Γ) sets capture Π3-part of T. In particular, the provably recursive functions of T are fFm j Fm(x) = minfy j [x; y] is m-dense(Γ)gg. Actually, one can generalize the above argument for infinite theories. One can also replace the base theory WKL0 with other systems, ∗ e.g., WKL0 or ACA0.
Recommended publications
  • A Note on Indicator-Functions
    proceedings of the american mathematical society Volume 39, Number 1, June 1973 A NOTE ON INDICATOR-FUNCTIONS J. MYHILL Abstract. A system has the existence-property for abstracts (existence property for numbers, disjunction-property) if when- ever V(1x)A(x), M(t) for some abstract (t) (M(n) for some numeral n; if whenever VAVB, VAor hß. (3x)A(x), A,Bare closed). We show that the existence-property for numbers and the disjunc- tion property are never provable in the system itself; more strongly, the (classically) recursive functions that encode these properties are not provably recursive functions of the system. It is however possible for a system (e.g., ZF+K=L) to prove the existence- property for abstracts for itself. In [1], I presented an intuitionistic form Z of Zermelo-Frankel set- theory (without choice and with weakened regularity) and proved for it the disfunction-property (if VAvB (closed), then YA or YB), the existence- property (if \-(3x)A(x) (closed), then h4(t) for a (closed) comprehension term t) and the existence-property for numerals (if Y(3x G m)A(x) (closed), then YA(n) for a numeral n). In the appendix to [1], I enquired if these results could be proved constructively; in particular if we could find primitive recursively from the proof of AwB whether it was A or B that was provable, and likewise in the other two cases. Discussion of this question is facilitated by introducing the notion of indicator-functions in the sense of the following Definition. Let T be a consistent theory which contains Heyting arithmetic (possibly by relativization of quantifiers).
    [Show full text]
  • Soft Set Theory First Results
    An Inlum~md computers & mathematics PERGAMON Computers and Mathematics with Applications 37 (1999) 19-31 Soft Set Theory First Results D. MOLODTSOV Computing Center of the R~msian Academy of Sciences 40 Vavilova Street, Moscow 117967, Russia Abstract--The soft set theory offers a general mathematical tool for dealing with uncertain, fuzzy, not clearly defined objects. The main purpose of this paper is to introduce the basic notions of the theory of soft sets, to present the first results of the theory, and to discuss some problems of the future. (~) 1999 Elsevier Science Ltd. All rights reserved. Keywords--Soft set, Soft function, Soft game, Soft equilibrium, Soft integral, Internal regular- ization. 1. INTRODUCTION To solve complicated problems in economics, engineering, and environment, we cannot success- fully use classical methods because of various uncertainties typical for those problems. There are three theories: theory of probability, theory of fuzzy sets, and the interval mathematics which we can consider as mathematical tools for dealing with uncertainties. But all these theories have their own difficulties. Theory of probabilities can deal only with stochastically stable phenomena. Without going into mathematical details, we can say, e.g., that for a stochastically stable phenomenon there should exist a limit of the sample mean #n in a long series of trials. The sample mean #n is defined by 1 n IZn = -- ~ Xi, n i=l where x~ is equal to 1 if the phenomenon occurs in the trial, and x~ is equal to 0 if the phenomenon does not occur. To test the existence of the limit, we must perform a large number of trials.
    [Show full text]
  • What Is Fuzzy Probability Theory?
    Foundations of Physics, Vol.30,No. 10, 2000 What Is Fuzzy Probability Theory? S. Gudder1 Received March 4, 1998; revised July 6, 2000 The article begins with a discussion of sets and fuzzy sets. It is observed that iden- tifying a set with its indicator function makes it clear that a fuzzy set is a direct and natural generalization of a set. Making this identification also provides sim- plified proofs of various relationships between sets. Connectives for fuzzy sets that generalize those for sets are defined. The fundamentals of ordinary probability theory are reviewed and these ideas are used to motivate fuzzy probability theory. Observables (fuzzy random variables) and their distributions are defined. Some applications of fuzzy probability theory to quantum mechanics and computer science are briefly considered. 1. INTRODUCTION What do we mean by fuzzy probability theory? Isn't probability theory already fuzzy? That is, probability theory does not give precise answers but only probabilities. The imprecision in probability theory comes from our incomplete knowledge of the system but the random variables (measure- ments) still have precise values. For example, when we flip a coin we have only a partial knowledge about the physical structure of the coin and the initial conditions of the flip. If our knowledge about the coin were com- plete, we could predict exactly whether the coin lands heads or tails. However, we still assume that after the coin lands, we can tell precisely whether it is heads or tails. In fuzzy probability theory, we also have an imprecision in our measurements, and random variables must be replaced by fuzzy random variables and events by fuzzy events.
    [Show full text]
  • Be Measurable Spaces. a Func- 1 1 Tion F : E G Is Measurable If F − (A) E Whenever a G
    2. Measurable functions and random variables 2.1. Measurable functions. Let (E; E) and (G; G) be measurable spaces. A func- 1 1 tion f : E G is measurable if f − (A) E whenever A G. Here f − (A) denotes the inverse!image of A by f 2 2 1 f − (A) = x E : f(x) A : f 2 2 g Usually G = R or G = [ ; ], in which case G is always taken to be the Borel σ-algebra. If E is a topological−∞ 1space and E = B(E), then a measurable function on E is called a Borel function. For any function f : E G, the inverse image preserves set operations ! 1 1 1 1 f − A = f − (A ); f − (G A) = E f − (A): i i n n i ! i [ [ 1 1 Therefore, the set f − (A) : A G is a σ-algebra on E and A G : f − (A) E is f 2 g 1 f ⊆ 2 g a σ-algebra on G. In particular, if G = σ(A) and f − (A) E whenever A A, then 1 2 2 A : f − (A) E is a σ-algebra containing A and hence G, so f is measurable. In the casef G = R, 2the gBorel σ-algebra is generated by intervals of the form ( ; y]; y R, so, to show that f : E R is Borel measurable, it suffices to show −∞that x 2E : f(x) y E for all y. ! f 2 ≤ g 2 1 If E is any topological space and f : E R is continuous, then f − (U) is open in E and hence measurable, whenever U is op!en in R; the open sets U generate B, so any continuous function is measurable.
    [Show full text]
  • Expectation and Functions of Random Variables
    POL 571: Expectation and Functions of Random Variables Kosuke Imai Department of Politics, Princeton University March 10, 2006 1 Expectation and Independence To gain further insights about the behavior of random variables, we first consider their expectation, which is also called mean value or expected value. The definition of expectation follows our intuition. Definition 1 Let X be a random variable and g be any function. 1. If X is discrete, then the expectation of g(X) is defined as, then X E[g(X)] = g(x)f(x), x∈X where f is the probability mass function of X and X is the support of X. 2. If X is continuous, then the expectation of g(X) is defined as, Z ∞ E[g(X)] = g(x)f(x) dx, −∞ where f is the probability density function of X. If E(X) = −∞ or E(X) = ∞ (i.e., E(|X|) = ∞), then we say the expectation E(X) does not exist. One sometimes write EX to emphasize that the expectation is taken with respect to a particular random variable X. For a continuous random variable, the expectation is sometimes written as, Z x E[g(X)] = g(x) d F (x). −∞ where F (x) is the distribution function of X. The expectation operator has inherits its properties from those of summation and integral. In particular, the following theorem shows that expectation preserves the inequality and is a linear operator. Theorem 1 (Expectation) Let X and Y be random variables with finite expectations. 1. If g(x) ≥ h(x) for all x ∈ R, then E[g(X)] ≥ E[h(X)].
    [Show full text]
  • More on Discrete Random Variables and Their Expectations
    MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2018 Lecture 6 MORE ON DISCRETE RANDOM VARIABLES AND THEIR EXPECTATIONS Contents 1. Comments on expected values 2. Expected values of some common random variables 3. Covariance and correlation 4. Indicator variables and the inclusion-exclusion formula 5. Conditional expectations 1COMMENTSON EXPECTED VALUES (a) Recall that E is well defined unless both sums and [X] x:x<0 xpX (x) E x:x>0 xpX (x) are infinite. Furthermore, [X] is well-defined and finite if and only if both sums are finite. This is the same as requiring! that ! E[|X|]= |x|pX (x) < ∞. x " Random variables that satisfy this condition are called integrable. (b) Noter that fo any random variable X, E[X2] is always well-defined (whether 2 finite or infinite), because all the terms in the sum x x pX (x) are nonneg- ative. If we have E[X2] < ∞,we say that X is square integrable. ! (c) Using the inequality |x| ≤ 1+x 2 ,wehave E [|X|] ≤ 1+E[X2],which shows that a square integrable random variable is always integrable. Simi- larly, for every positive integer r,ifE [|X|r] is finite then it is also finite for every l<r(fill details). 1 Exercise 1. Recall that the r-the central moment of a random variable X is E[(X − E[X])r].Showthat if the r -th central moment of an almost surely non-negative random variable X is finite, then its l-th central moment is also finite for every l<r. (d) Because of the formula var(X)=E[X2] − (E[X])2,wesee that: (i) if X is square integrable, the variance is finite; (ii) if X is integrable, but not square integrable, the variance is infinite; (iii) if X is not integrable, the variance is undefined.
    [Show full text]
  • POST-OLYMPIAD PROBLEMS (1) Evaluate ∫ Cos2k(X)Dx Using Combinatorics. (2) (Putnam) Suppose That F : R → R Is Differentiable
    POST-OLYMPIAD PROBLEMS MARK SELLKE R 2π 2k (1) Evaluate 0 cos (x)dx using combinatorics. (2) (Putnam) Suppose that f : R ! R is differentiable, and that for all q 2 Q we have f 0(q) = f(q). Need f(x) = cex for some c? 1 1 (3) Show that the random sum ±1 ± 2 ± 3 ::: defines a conditionally convergent series with probability 1. 2 (4) (Jacob Tsimerman) Does there exist a smooth function f : R ! R with a unique critical point at 0, such that 0 is a local minimum but not a global minimum? (5) Prove that if f 2 C1([0; 1]) and for each x 2 [0; 1] there is n = n(x) with f (n) = 0, then f is a polynomial. 2 2 (6) (Stanford Math Qual 2018) Prove that the operator χ[a0;b0]Fχ[a1;b1] from L ! L is always compact. Here if S is a set, χS is multiplication by the indicator function 1S. And F is the Fourier transform. (7) Let n ≥ 4 be a positive integer. Show there are no non-trivial solutions in entire functions to f(z)n + g(z)n = h(z)n. (8) Let x 2 Sn be a random point on the unit n-sphere for large n. Give first-order asymptotics for the largest coordinate of x. (9) Show that almost sure convergence is not equivalent to convergence in any topology on random variables. P1 xn (10) (Noam Elkies on MO) Show that n=0 (n!)α is positive for any α 2 (0; 1) and any real x.
    [Show full text]
  • 10.0 Lesson Plan
    10.0 Lesson Plan 1 Answer Questions • Robust Estimators • Maximum Likelihood Estimators • 10.1 Robust Estimators Previously, we claimed to like estimators that are unbiased, have minimum variance, and/or have minimum mean squared error. Typically, one cannot achieve all of these properties with the same estimator. 2 An estimator may have good properties for one distribution, but not for n another. We saw that n−1 Z, for Z the sample maximum, was excellent in estimating θ for a Unif(0,θ) distribution. But it would not be excellent for estimating θ when, say, the density function looks like a triangle supported on [0,θ]. A robust estimator is one that works well across many families of distributions. In particular, it works well when there may be outliers in the data. The 10% trimmed mean is a robust estimator of the population mean. It discards the 5% largest and 5% smallest observations, and averages the rest. (Obviously, one could trim by some fraction other than 10%, but this is a commonly-used value.) Surveyors distinguish errors from blunders. Errors are measurement jitter 3 attributable to chance effects, and are approximately Gaussian. Blunders occur when the guy with the theodolite is standing on the wrong hill. A trimmed mean throws out the blunders and averages the good data. If all the data are good, one has lost some sample size. But in exchange, you are protected from the corrosive effect of outliers. 10.2 Strategies for Finding Estimators Some economists focus on the method of moments. This is a terrible procedure—its only virtue is that it is easy.
    [Show full text]
  • Overview of Measure Theory
    Overview of Measure Theory January 14, 2013 In this section, we will give a brief overview of measure theory, which leads to a general notion of an integral called the Lebesgue integral. Integrals, as we saw before, are important in probability theory since the notion of expectation or average value is an integral. The ideas presented in this theory are fairly general, but their utility will not be immediately visible. However, a general note is that we are trying to quantify how \small" or how \large" a set is. Recall that if a bounded function has only finitely many discontinuities, it is still Riemann-integrable. The Dirichlet function was an example of a function with uncountably many discontinuities, and it failed to be Riemann-integrable. Is there a notion of \smallness" such that if a function is bounded except for a small set, then we can compute its integral? For example, we could say a set is \small" on the real line if it has at most countably infinitely many elements. Unfortunately, this is not enough to deal with the integral of the Dirichlet function. So we need a notion adequate to deal with the integral of functions such as the Dirichlet function. We will deal with such a notion of integral in this section. It allows us to prove a fairly general theorem called the \Dominated Convergence Theorem", which does not hold for the Riemann integral, and is useful for some of the results in our course. The general outline is the following - first, we deal with a notion of the class of measurable sets, which will restrict the sets whose size we can speak of.
    [Show full text]
  • 6.436J / 15.085J Fundamentals of Probability, Lecture 4: Random
    MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2018 Lecture 4 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete random variables 4. Continuous random variables A. Continuity B. The Cantor set, an unusual random variable, and a singular measure 1 RANDOM VARIABLES AND MEASURABLE FUNCTIONS Loosely speaking a random variable provides us with a numerical value, depend- ing on the outcome of an experiment. More precisely, a random variable can be viewed as a function from the sample space to the real numbers, and we will use the notation X(!) to denote the numerical value of a random variable X, when the outcome of the experiment is some particular !. We may be interested in the probability that the outcome of the experiment is such that X is no larger than some c, i.e., that the outcome belongs to the set ! X(!) c . Of course, in { | ≤ } order to have a probability assigned to that set, we need to make sure that it is -measurable. This motivates Definition 1 below. F Example 1. Consider a sequence of five consecutive coin tosses. An appropriate sample space is Ω = 0, 1 n , where “1” stands for heads and “0” for tails. Let be the col- { } F lection of all subsets of Ω , and suppose that a probability measure P has been assigned to ( Ω, ). We are interested in the number of heads obtained in this experiment. This F quantity can be described by the function X : Ω R, defined by → X(!1,...,! ) = !1 + + ! . n ··· n 1 With this definition, the set ! X(!) < 4 is just the event that there were fewer than { | } 4 heads overall, belongs to the ˙-field , and therefore has a well-defined probability.
    [Show full text]
  • 18.175: Lecture 3 .1In Integration
    Random variables Integration Expectation 18.175: Lecture 3 Integration Scott Sheffield MIT 18.175 Lecture 3 Random variables Integration Expectation Outline Random variables Integration Expectation 18.175 Lecture 3 Random variables Integration Expectation Outline Random variables Integration Expectation 18.175 Lecture 3 I σ-algebra is collection of subsets closed under complementation and countable unions. Call (Ω; F) a measure space. I Measure is function µ : F! R satisfying µ(A) ≥ µ(;) = 0 P for all A 2 F and countable additivity: µ([i Ai ) = i µ(Ai ) for disjoint Ai . I Measure µ is probability measure if µ(Ω) = 1. I The Borel σ-algebra B on a topological space is the smallest σ-algebra containing all open sets. Random variables Integration Expectation Recall definitions I Probability space is triple (Ω; F; P) where Ω is sample space, F is set of events (the σ-algebra) and P : F! [0; 1] is the probability function. 18.175 Lecture 3 I Measure is function µ : F! R satisfying µ(A) ≥ µ(;) = 0 P for all A 2 F and countable additivity: µ([i Ai ) = i µ(Ai ) for disjoint Ai . I Measure µ is probability measure if µ(Ω) = 1. I The Borel σ-algebra B on a topological space is the smallest σ-algebra containing all open sets. Random variables Integration Expectation Recall definitions I Probability space is triple (Ω; F; P) where Ω is sample space, F is set of events (the σ-algebra) and P : F! [0; 1] is the probability function. I σ-algebra is collection of subsets closed under complementation and countable unions.
    [Show full text]
  • Random Variables and Expectation
    Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 5: Random variables and expectation Relevant textbook passages: Pitman [5]: Sections 3.1–3.2 Larsen–Marx [4]: Sections 3.3–3.5 5.1 Random variables 5.1.1 Definition A random variable on a probability space (S, E,P ) is a real-valued function on S which has the property that for every interval I ⊂ R the inverse image of I is an event. Note that when the collection E of events consists of all subsets of S, then the requirement that inverse images of intervals be events is automatically satisfied. 5.1.2 Remark An interpretation of random variables used by engineers is that they represent measurements on the state of a system. See, e.g., Robert Gray [3]. There is another definition of random variable that is quite common, especially in electrical engineering. 5.1.3 Definition (Another kind of random variable) Given a set A of symbols or letters, called the alphabet, a random variable is defined to be a function from S into A. While we could enumerate the symbols in the alphabet and treat the random variable as a real-valued function, the arithmetic operations have no significance: what letter is the sumof the letters A and B? Traditionally, probabilists and statisticians use upper-case Latin letters near the end of the alphabet to denote random variables. This has confused generations of students, who have trouble thinking of random variables as functions. For the sake of tradition, and so that you get used to it, we follow suit.
    [Show full text]