AXIOMS of PROBABILITY OCT 1, 2018 Sets and Operations on Sets

Total Page:16

File Type:pdf, Size:1020Kb

AXIOMS of PROBABILITY OCT 1, 2018 Sets and Operations on Sets AXIOMS OF PROBABILITY OCT 1, 2018 Sets and operations on sets. Appendix B Review Venn diagrams and set operations: (i) union A [ B, (ii) intersection A \ B (also written as AB), (iii) complement Ac, (iv) difference AnB. Problem. Let A; B; C be subsets of Ω. Represent the following in set notation. (1) Set of elements that are in each of the three subsets = A \ B \ C. (2) Set of elements that are in A but neither in B nor in C = A \ Bc \ Cc. (3) Set of elements that are in at least one of the sets A or B = A [ B. (4) Set of elements that re in both A and B, but not in C = A \ B \ Cc. (5) Set of elements that are in A, but not in B or C or both = An(B [ C). Axiomatic definition of a probability measure. Let Ω (pronounced `omega') be a set called sample space. Elements of Ω are called outcomes while subsets of Ω are called events. Let F be the power set (i.e., set of all subsets of Ω) of Ω. Then a probability, or probability measure, is a function from F to [0; 1] satisfying the following conditions: • P (Ω) = 1. P (;) = 0. • Countable additivity: If A1;A2;::: is a sequence of pairwise disjoint (i.e. Ai \ Aj = ; for all i 6= j) events then 1 1 X P ([i=1Ai) = P (Ai): i=1 In other words, with every event E ⊆ Ω, a probability measure associates a number 0 ≤ P (E) ≤ 1 which is the probability that we observe the event E happening. The triplet (Ω; F;P ) is called a probability space. Remark 1. More generally F is taken to be a σ-field (pronounced ‘sigma-field’), a concept from measure theory. But you don't need to know that right now. Some consequences of the definition. • Finite additivity: If A1;:::;Ak is a finite pairwise disjoint collection of events, then k P [i=1Ai = P (A1) + ::: + P (Ak): This follows from countable additivity by taking Ak+1 = Ak+2 = ::: = ;. • Complements: If A is an event, its complement Ac is the event Ac = ΩnA. Clearly A and Ac are disjoint. Thus, by finite additivity, P (A) + P (Ac) = P (A \ Ac) = P (Ω) = 1. Thus P (Ac) = 1 − P (A): Equally likely outcomes. Suppose the sample space Ω is finite and has N ele- ments f!1;:::;!N g. Say that all outcomes are equally likely if P (!i) := P (f!ig) = 1=N for all i. In that case, it is very easy to find the probability measure. If E is 1 2 an event, then E is a subset of Ω. If #E is the number of elements in E then, by finite additivity, #E P (E) = : N Check that obviously P (Ω) = 1 and P (;) = 0. Example 1. Toss a coin twice. Then Ω = fHH;TT;HT;THg, and 1 P fHHg = P fTT g = P fTHg = P fHT g = : 4 For example, the event A = fHH; T T g, i.e., the event that both tosses are the same, has probability P (A) = 1=2, under the assumption of equally likely outcomes. Example 2. Roll a die twice. Here Ω is the set of 36 possible outcomes f(1; 1); (1; 2); (1; 3);::: (6; 5); (6; 6)g: Let E be the event that the sum of the two rolls is 5. Then E consists of 4 sample points: (1; 4); (2; 3); (3; 2); (4; 1). Hence P (E) = 4=36 = 1=9, under the assumption of equally likely outcomes. Example 3. Sampling with replacement. Imagine an urn with N ≥ 2 balls la- beled by f1; 2;:::;Ng. Pick one ball without looking, note its label, and put it back. Repeat this a total of k times. The k labels picked form a k-tuple (a1; a2; : : : ; ak). This is one random outcome. The sample space Ω is the set of all k-tuples with entries in f1; 2;:::;Ng. Clearly Ω is finite and has N k elements. We say that the sampling is random, or, more correctly uniformly at random, if every outcome in Ω is equally likely. Hence the probability of every tuple is exactly 1=N k, under the assumption of equally likely outcomes. What is the probability that every label sampled is either 1 or 2? Here the event k E := f(!1;:::;!k): !i is 1 or 2 for each ig. Clearly there are #E = 2 . Hence P (E) = 2k=N k = (2=N)k, under the assumption of equally likely outcomes. Example 4. Ordered sampling without replacement. In sampling without replacement, we still have an urn with N ≥ 2 balls labeled by f1; 2;:::;Ng. We pick balls successively and note its label. However, we do not put it back. Hence Ω is the set of arrangements of k labels taken from f1; 2;:::;Ng. Thus #Ω = (n)k. A random ordered sample without replacement means each of these (n)k outcomes are equally likely. What is the probability that every label sampled is either 1 or 2? It is zero if k > 2, since after the first two labels are 1 or 2, the rest cannot be either. Example 5. (Unordered) sampling without replacement. Take an urn with N ≥ 2 balls labeled by f1; 2;:::;Ng. By a random sample without replace- n ment we mean that pick any subset of k balls equally likely. Thus #Ω = k ..
Recommended publications
  • Calibrating Determinacy Strength in Levels of the Borel Hierarchy
    CALIBRATING DETERMINACY STRENGTH IN LEVELS OF THE BOREL HIERARCHY SHERWOOD J. HACHTMAN Abstract. We analyze the set-theoretic strength of determinacy for levels of the Borel 0 hierarchy of the form Σ1+α+3, for α < !1. Well-known results of H. Friedman and D.A. Martin have shown this determinacy to require α+1 iterations of the Power Set Axiom, but we ask what additional ambient set theory is strictly necessary. To this end, we isolate a family of Π1-reflection principles, Π1-RAPα, whose consistency strength corresponds 0 CK exactly to that of Σ1+α+3-Determinacy, for α < !1 . This yields a characterization of the levels of L by or at which winning strategies in these games must be constructed. When α = 0, we have the following concise result: the least θ so that all winning strategies 0 in Σ4 games belong to Lθ+1 is the least so that Lθ j= \P(!) exists + all wellfounded trees are ranked". x1. Introduction. Given a set A ⊆ !! of sequences of natural numbers, consider a game, G(A), where two players, I and II, take turns picking elements of a sequence hx0; x1; x2;::: i of naturals. Player I wins the game if the sequence obtained belongs to A; otherwise, II wins. For a collection Γ of subsets of !!, Γ determinacy, which we abbreviate Γ-DET, is the statement that for every A 2 Γ, one of the players has a winning strategy in G(A). It is a much-studied phenomenon that Γ -DET has mathematical strength: the bigger the pointclass Γ, the stronger the theory required to prove Γ -DET.
    [Show full text]
  • A Proof of Cantor's Theorem
    Cantor’s Theorem Joe Roussos 1 Preliminary ideas Two sets have the same number of elements (are equinumerous, or have the same cardinality) iff there is a bijection between the two sets. Mappings: A mapping, or function, is a rule that associates elements of one set with elements of another set. We write this f : X ! Y , f is called the function/mapping, the set X is called the domain, and Y is called the codomain. We specify what the rule is by writing f(x) = y or f : x 7! y. e.g. X = f1; 2; 3g;Y = f2; 4; 6g, the map f(x) = 2x associates each element x 2 X with the element in Y that is double it. A bijection is a mapping that is injective and surjective.1 • Injective (one-to-one): A function is injective if it takes each element of the do- main onto at most one element of the codomain. It never maps more than one element in the domain onto the same element in the codomain. Formally, if f is a function between set X and set Y , then f is injective iff 8a; b 2 X; f(a) = f(b) ! a = b • Surjective (onto): A function is surjective if it maps something onto every element of the codomain. It can map more than one thing onto the same element in the codomain, but it needs to hit everything in the codomain. Formally, if f is a function between set X and set Y , then f is surjective iff 8y 2 Y; 9x 2 X; f(x) = y Figure 1: Injective map.
    [Show full text]
  • Probability and Statistics Lecture Notes
    Probability and Statistics Lecture Notes Antonio Jiménez-Martínez Chapter 1 Probability spaces In this chapter we introduce the theoretical structures that will allow us to assign proba- bilities in a wide range of probability problems. 1.1. Examples of random phenomena Science attempts to formulate general laws on the basis of observation and experiment. The simplest and most used scheme of such laws is: if a set of conditions B is satisfied =) event A occurs. Examples of such laws are the law of gravity, the law of conservation of mass, and many other instances in chemistry, physics, biology... If event A occurs inevitably whenever the set of conditions B is satisfied, we say that A is certain or sure (under the set of conditions B). If A can never occur whenever B is satisfied, we say that A is impossible (under the set of conditions B). If A may or may not occur whenever B is satisfied, then A is said to be a random phenomenon. Random phenomena is our subject matter. Unlike certain and impossible events, the presence of randomness implies that the set of conditions B do not reflect all the necessary and sufficient conditions for the event A to occur. It might seem them impossible to make any worthwhile statements about random phenomena. However, experience has shown that many random phenomena exhibit a statistical regularity that makes them subject to study. For such random phenomena it is possible to estimate the chance of occurrence of the random event. This estimate can be obtained from laws, called probabilistic or stochastic, with the form: if a set of conditions B is satisfied event A occurs m times =) repeatedly n times out of the n repetitions.
    [Show full text]
  • The Probability Set-Up.Pdf
    CHAPTER 2 The probability set-up 2.1. Basic theory of probability We will have a sample space, denoted by S (sometimes Ω) that consists of all possible outcomes. For example, if we roll two dice, the sample space would be all possible pairs made up of the numbers one through six. An event is a subset of S. Another example is to toss a coin 2 times, and let S = fHH;HT;TH;TT g; or to let S be the possible orders in which 5 horses nish in a horse race; or S the possible prices of some stock at closing time today; or S = [0; 1); the age at which someone dies; or S the points in a circle, the possible places a dart can hit. We should also keep in mind that the same setting can be described using dierent sample set. For example, in two solutions in Example 1.30 we used two dierent sample sets. 2.1.1. Sets. We start by describing elementary operations on sets. By a set we mean a collection of distinct objects called elements of the set, and we consider a set as an object in its own right. Set operations Suppose S is a set. We say that A ⊂ S, that is, A is a subset of S if every element in A is contained in S; A [ B is the union of sets A ⊂ S and B ⊂ S and denotes the points of S that are in A or B or both; A \ B is the intersection of sets A ⊂ S and B ⊂ S and is the set of points that are in both A and B; ; denotes the empty set; Ac is the complement of A, that is, the points in S that are not in A.
    [Show full text]
  • TOPOLOGY and ITS APPLICATIONS the Number of Complements in The
    TOPOLOGY AND ITS APPLICATIONS ELSEVIER Topology and its Applications 55 (1994) 101-125 The number of complements in the lattice of topologies on a fixed set Stephen Watson Department of Mathematics, York Uniuersity, 4700 Keele Street, North York, Ont., Canada M3J IP3 (Received 3 May 1989) (Revised 14 November 1989 and 2 June 1992) Abstract In 1936, Birkhoff ordered the family of all topologies on a set by inclusion and obtained a lattice with 1 and 0. The study of this lattice ought to be a basic pursuit both in combinatorial set theory and in general topology. In this paper, we study the nature of complementation in this lattice. We say that topologies 7 and (T are complementary if and only if 7 A c = 0 and 7 V (T = 1. For simplicity, we call any topology other than the discrete and the indiscrete a proper topology. Hartmanis showed in 1958 that any proper topology on a finite set of size at least 3 has at least two complements. Gaifman showed in 1961 that any proper topology on a countable set has at least two complements. In 1965, Steiner showed that any topology has a complement. The question of the number of distinct complements a topology on a set must possess was first raised by Berri in 1964 who asked if every proper topology on an infinite set must have at least two complements. In 1969, Schnare showed that any proper topology on a set of infinite cardinality K has at least K distinct complements and at most 2” many distinct complements.
    [Show full text]
  • Determinacy in Linear Rational Expectations Models
    Journal of Mathematical Economics 40 (2004) 815–830 Determinacy in linear rational expectations models Stéphane Gauthier∗ CREST, Laboratoire de Macroéconomie (Timbre J-360), 15 bd Gabriel Péri, 92245 Malakoff Cedex, France Received 15 February 2002; received in revised form 5 June 2003; accepted 17 July 2003 Available online 21 January 2004 Abstract The purpose of this paper is to assess the relevance of rational expectations solutions to the class of linear univariate models where both the number of leads in expectations and the number of lags in predetermined variables are arbitrary. It recommends to rule out all the solutions that would fail to be locally unique, or equivalently, locally determinate. So far, this determinacy criterion has been applied to particular solutions, in general some steady state or periodic cycle. However solutions to linear models with rational expectations typically do not conform to such simple dynamic patterns but express instead the current state of the economic system as a linear difference equation of lagged states. The innovation of this paper is to apply the determinacy criterion to the sets of coefficients of these linear difference equations. Its main result shows that only one set of such coefficients, or the corresponding solution, is locally determinate. This solution is commonly referred to as the fundamental one in the literature. In particular, in the saddle point configuration, it coincides with the saddle stable (pure forward) equilibrium trajectory. © 2004 Published by Elsevier B.V. JEL classification: C32; E32 Keywords: Rational expectations; Selection; Determinacy; Saddle point property 1. Introduction The rational expectations hypothesis is commonly justified by the fact that individual forecasts are based on the relevant theory of the economic system.
    [Show full text]
  • 180: Counting Techniques
    180: Counting Techniques In the following exercise we demonstrate the use of a few fundamental counting principles, namely the addition, multiplication, complementary, and inclusion-exclusion principles. While none of the principles are particular complicated in their own right, it does take some practice to become familiar with them, and recognise when they are applicable. I have attempted to indicate where alternate approaches are possible (and reasonable). Problem: Assume n ≥ 2 and m ≥ 1. Count the number of functions f :[n] ! [m] (i) in total. (ii) such that f(1) = 1 or f(2) = 1. (iii) with minx2[n] f(x) ≤ 5. (iv) such that f(1) ≥ f(2). (v) that are strictly increasing; that is, whenever x < y, f(x) < f(y). (vi) that are one-to-one (injective). (vii) that are onto (surjective). (viii) that are bijections. (ix) such that f(x) + f(y) is even for every x; y 2 [n]. (x) with maxx2[n] f(x) = minx2[n] f(x) + 1. Solution: (i) A function f :[n] ! [m] assigns for every integer 1 ≤ x ≤ n an integer 1 ≤ f(x) ≤ m. For each integer x, we have m options. As we make n such choices (independently), the total number of functions is m × m × : : : m = mn. (ii) (We assume n ≥ 2.) Let A1 be the set of functions with f(1) = 1, and A2 the set of functions with f(2) = 1. Then A1 [ A2 represents those functions with f(1) = 1 or f(2) = 1, which is precisely what we need to count. We have jA1 [ A2j = jA1j + jA2j − jA1 \ A2j.
    [Show full text]
  • 1 Probabilities
    1 Probabilities 1.1 Experiments with randomness We will use the term experiment in a very general way to refer to some process that produces a random outcome. Examples: (Ask class for some first) Here are some discrete examples: • roll a die • flip a coin • flip a coin until we get heads And here are some continuous examples: • height of a U of A student • random number in [0, 1] • the time it takes until a radioactive substance undergoes a decay These examples share the following common features: There is a proce- dure or natural phenomena called the experiment. It has a set of possible outcomes. There is a way to assign probabilities to sets of possible outcomes. We will call this a probability measure. 1.2 Outcomes and events Definition 1. An experiment is a well defined procedure or sequence of procedures that produces an outcome. The set of possible outcomes is called the sample space. We will typically denote an individual outcome by ω and the sample space by Ω. Definition 2. An event is a subset of the sample space. This definition will be changed when we come to the definition ofa σ-field. The next thing to define is a probability measure. Before we can do this properly we need some more structure, so for now we just make an informal definition. A probability measure is a function on the collection of events 1 that assign a number between 0 and 1 to each event and satisfies certain properties. NB: A probability measure is not a function on Ω.
    [Show full text]
  • Propensities and Probabilities
    ARTICLE IN PRESS Studies in History and Philosophy of Modern Physics 38 (2007) 593–625 www.elsevier.com/locate/shpsb Propensities and probabilities Nuel Belnap 1028-A Cathedral of Learning, University of Pittsburgh, Pittsburgh, PA 15260, USA Received 19 May 2006; accepted 6 September 2006 Abstract Popper’s introduction of ‘‘propensity’’ was intended to provide a solid conceptual foundation for objective single-case probabilities. By considering the partly opposed contributions of Humphreys and Miller and Salmon, it is argued that when properly understood, propensities can in fact be understood as objective single-case causal probabilities of transitions between concrete events. The chief claim is that propensities are well-explicated by describing how they fit into the existing formal theory of branching space-times, which is simultaneously indeterministic and causal. Several problematic examples, some commonsense and some quantum-mechanical, are used to make clear the advantages of invoking branching space-times theory in coming to understand propensities. r 2007 Elsevier Ltd. All rights reserved. Keywords: Propensities; Probabilities; Space-times; Originating causes; Indeterminism; Branching histories 1. Introduction You are flipping a fair coin fairly. You ascribe a probability to a single case by asserting The probability that heads will occur on this very next flip is about 50%. ð1Þ The rough idea of a single-case probability seems clear enough when one is told that the contrast is with either generalizations or frequencies attributed to populations asserted while you are flipping a fair coin fairly, such as In the long run; the probability of heads occurring among flips is about 50%. ð2Þ E-mail address: [email protected] 1355-2198/$ - see front matter r 2007 Elsevier Ltd.
    [Show full text]
  • Power Sets Math 12, Veritas Prep
    Power Sets Math 12, Veritas Prep. The power set P (X) of a set X is the set of all subsets of X. Defined formally, P (X) = fK j K ⊆ Xg (i.e., the set of all K such that K is a subset of X). Let me give an example. Suppose it's Thursday night, and I want to go to a movie. And suppose that I know that each of the upper-campus Latin teachers|Sullivan, Joyner, and Pagani|are free that night. So I consider asking one of them if they want to join me. Conceivably, then: • I could go to the movie by myself (i.e., I could bring along none of the Latin faculty) • I could go to the movie with Sullivan • I could go to the movie with Joyner • I could go to the movie with Pagani • I could go to the movie with Sullivan and Joyner • I could go to the movie with Sullivan and Pagani • I could go to the movie with Pagani and Joyner • or I could go to the movie with all three of them Put differently, my choices for whom to go to the movie with make up the following set: fg; fSullivang; fJoynerg; fPaganig; fSullivan, Joynerg; fSullivan, Paganig; fPagani, Joynerg; fSullivan, Joyner, Paganig Or if I use my notation for the empty set: ;; fSullivang; fJoynerg; fPaganig; fSullivan, Joynerg; fSullivan, Paganig; fPagani, Joynerg; fSullivan, Joyner, Paganig This set|the set of whom I might bring to the movie|is the power set of the set of Latin faculty. It is the set of all the possible subsets of the set of Latin faculty.
    [Show full text]
  • Set Difference and Symmetric Difference of Fuzzy Sets
    Preliminaries Symmetric Dierence Set Dierence and Symmetric Dierence of Fuzzy Sets N.R. Vemuri A.S. Hareesh M.S. Srinath Department of Mathematics Indian Institute of Technology, Hyderabad and Department of Mathematics and Computer Science Sri Sathya Sai Institute of Higher Learning, India Fuzzy Sets Theory and Applications 2014, Liptovský Ján, Slovak Republic Vemuri, Sai Hareesh & Srinath Symmetric Dierence Preliminaries Introduction Symmetric Dierence Earlier work Outline 1 Preliminaries Introduction Earlier work 2 Symmetric Dierence Denition Examples Properties Applications Future Work References Vemuri, Sai Hareesh & Srinath Symmetric Dierence Preliminaries Introduction Symmetric Dierence Earlier work Classical set theory Set operations Union- [ Intersection - \ Complement - c Dierence -n Symmetric dierence - ∆ .... Vemuri, Sai Hareesh & Srinath Symmetric Dierence Preliminaries Introduction Symmetric Dierence Earlier work Classical set theory Set operations Union- [ Intersection - \ Complement - c Dierence -n Symmetric dierence - ∆ .... Vemuri, Sai Hareesh & Srinath Symmetric Dierence Preliminaries Introduction Symmetric Dierence Earlier work Classical set theory Set operations Union- [ Intersection - \ Complement - c Dierence -n Symmetric dierence - ∆ .... Vemuri, Sai Hareesh & Srinath Symmetric Dierence Preliminaries Introduction Symmetric Dierence Earlier work Classical set theory Set operations Union- [ Intersection - \ Complement - c Dierence -n Symmetric dierence - ∆ .... Vemuri, Sai Hareesh & Srinath Symmetric Dierence Preliminaries
    [Show full text]
  • The Axiom of Choice and Its Implications
    THE AXIOM OF CHOICE AND ITS IMPLICATIONS KEVIN BARNUM Abstract. In this paper we will look at the Axiom of Choice and some of the various implications it has. These implications include a number of equivalent statements, and also some less accepted ideas. The proofs discussed will give us an idea of why the Axiom of Choice is so powerful, but also so controversial. Contents 1. Introduction 1 2. The Axiom of Choice and Its Equivalents 1 2.1. The Axiom of Choice and its Well-known Equivalents 1 2.2. Some Other Less Well-known Equivalents of the Axiom of Choice 3 3. Applications of the Axiom of Choice 5 3.1. Equivalence Between The Axiom of Choice and the Claim that Every Vector Space has a Basis 5 3.2. Some More Applications of the Axiom of Choice 6 4. Controversial Results 10 Acknowledgments 11 References 11 1. Introduction The Axiom of Choice states that for any family of nonempty disjoint sets, there exists a set that consists of exactly one element from each element of the family. It seems strange at first that such an innocuous sounding idea can be so powerful and controversial, but it certainly is both. To understand why, we will start by looking at some statements that are equivalent to the axiom of choice. Many of these equivalences are very useful, and we devote much time to one, namely, that every vector space has a basis. We go on from there to see a few more applications of the Axiom of Choice and its equivalents, and finish by looking at some of the reasons why the Axiom of Choice is so controversial.
    [Show full text]