Lecture Notes for the Introduction to Probability Course

Total Page:16

File Type:pdf, Size:1020Kb

Lecture Notes for the Introduction to Probability Course Lecture Notes for the Introduction to Probability Course Vladislav Kargin December 5, 2020 Contents 1 Combinatorial Probability and Basic Laws of Probabilistic Inference 4 1.1 What is randomness and what is probability?. 4 1.1.1 The classical definition of probability . 4 1.1.2 Kolmogorov’s Axiomatic Probability . 9 1.1.3 More about set theory and indicator functions . 12 1.2 Discrete Probability Models. 16 1.3 Sample-point method (Combinatorial probability) . 19 1.4 Conditional Probability and Independence . 29 1.4.1 Conditional probability . 30 1.4.2 Independence . 32 1.4.3 Law of total probability . 34 1.5 Bayes’ formula . 39 2 Discrete random variables and probability distributions. 44 2.1 Pmf and cdf . 44 2.2 Expected value . 49 2.2.1 Definition . 49 2.2.2 Expected value for a function of a r.v. 51 2.2.3 Properties of the expected value . 53 2.3 Variance and standard deviation . 55 2.3.1 Definition . 55 2.3.2 Properties. 57 1 2.4 The zoo of discrete random variables . 58 2.4.1 Binomial . 59 2.4.2 Geometric. 63 2.4.3 Negative binomial . 65 2.4.4 Hypergeometric . 68 2.4.5 Poisson . 70 2.5 Moments and moment-generating function . 75 2.6 Markov’s and Chebyshev’s inequalities . 80 3 Chapter 4: Continuous random variables. 84 3.1 Cdf and pdf . 84 3.2 Expected value and variance . 89 3.3 Quantiles . 91 3.4 The Uniform random variable . 92 3.5 The normal (or Gaussian) random variable. 94 3.6 Exponential and Gamma distributions . 101 3.7 Beta distribution . 107 4 Multivariate Distributions 110 4.1 Discrete random variables . 111 4.1.1 Joint and marginal pmf . 111 4.1.2 Conditional pmf and conditional expectation . 114 4.2 Dependence between random variables . 119 4.2.1 Independent random variables. 119 4.2.2 Covariance . 122 4.2.3 Correlation coefficient . 126 4.3 Continuous random variables . 127 4.3.1 Joint cdf and pdf . 127 4.3.2 Calculating probabilities using joint pdf . 128 4.3.3 Marginal densities . 130 4.3.4 Conditional density (pdf). 134 4.3.5 Conditional Expectations. 136 4.3.6 Independence for continuous random variables . 137 2 4.3.7 Expectations of functions and covariance for continuous r.v.’s . 140 4.3.8 Variance and conditional variance . 142 4.4 The multinomial distribution . 143 4.5 The multivariate normal distribution . 145 5 Functions of Random Variables 147 5.1 The method of cumulative distribution functions . 148 5.1.1 Functions of one random variable . 148 5.1.2 Functions of several random variables . 150 5.2 Random variable generation. 152 5.3 The pdf transformation method . 152 5.3.1 A function of one random variable . 152 5.3.2 Functions of several variables . 154 5.4 Method of moment-generating functions . 159 5.5 Order Statistics . 160 6 Sampling distributions, Law of Large Numbers, and Cen- tral Limit Theorem 168 6.1 Sampling distributions . 168 6.2 Law of Large Numbers (LLN) . 175 6.3 The Central Limit Theorem. 176 3 Chapter 1 Combinatorial Probability and Basic Laws of Probabilistic Inference 1.1 What is randomness and what is probability? “Probability is the most important concept in modern science, especially as nobody has the slightest notion what it means.” Bertrand Russell 1.1.1 The classical definition of probability The probability theory grew up from the study of games of chance, and the definition of probability in times of French mathematician Laplace (who taught mathematics to Napoleon) was very simple. We start by considering this definition and move to a more modern definition due to Kolmogorov. 4 Definition 1.1.1 (Laplace’s definition). If each case is equally probable, then the probability of a random event comprised of several cases is equal to the number of favorable cases divided by the number of all possible cases. From the point of view of this definition, probability is the art of counting all possible and all favorable cases. We will practice this art in the first 5% of the course. Example 1.1.2. Roll a fair die. What is the probability that you get an even number? This is easy. However, it might become a bit more difficult to define what are equally probable outcomes if you consider more complicated examples. Example 1.1.3. Suppose that we throw two fair dice. Is a throw of 11 as likely as a throw of 12? ( A throw is the sum of the numbers on the dice.) Leibniz (one of the inventors of calculus) wrote in a letter: “... For example, with two dice, it is equally likely to throw twelve points, than to throw eleven; because one or the other can be done in only one manner.” Leibniz makes a mistake here. In order to see it, look at this super-simple example. Example 1.1.4. Suppose we toss two fair coins. What is the probability that we observe a head and a tail? Leibniz would presumably answered 1/3. There are only 3 possibilities: Head and Head, Head and Tail, Tail and Tail. All possibilities are equally probable, so every one of them should have probability 1/3. However, if we do a large number of experiments with coin tossing then we find that the relative frequency in a large number of experiments isclose to 1/2 not to 1/3. Maybe these possibilities are not equally probable, after all. Indeed, the coins are distinguishable and the list of possible outcomes is in fact HH = “H on first coin, H on the second coin”, HT, TH anTT.If we assume that these 4 outcomes are equally probable, then there are two favorable outcomes: HT and TH and, by Laplace’s definition, the proba- bility is 2/4 = 1/2. This agrees with experiment eventually justifying our assumption. 5 The assumption that the coins are distinguishable and the four possible outcomes are equiprobable works in the real world which we observe every day. However, if we consider less usual objects, for example, objects in quantum mechanics, then we encounter objects for which this assumption is violated. In particular, elementary particles called bosons, – for example, photons are bosons, – are indistinguishable. So if we setup an experiment in which we use polarizations of two photons as coins, then possible outcomes in this experiment will be HH, HT and TT, since we are not able to say which photon is which. Moreover, these three outcomes are equiprobable. For this experiment, Leibnitz’ prediction is correct and the probability of HT is 1/3. The probability rules based on this set of assumptions are called the Bose- Einstein statistics. Note that Laplace’s definition is OK in this situation. It simply happens that the set of equiprobable outcomes is different. Another set of probability rules that arise in quantum mechanics is Fermi-Dirac statistics for fermion particles. If the coins behaved like electron spins, then only HT is a possible outcome and the probability to observe it is 1. We will use the standard assumption for coins and similar objects in this course and we will do many calculations based on Laplace’s definition of probability. Why do we need something else besides the Laplace definition? Even in usual circumstances, there are situations in which Laplace’s definition is not sufficient. This is especially relevant when wehavean uncountable infinity of possible outcomes. Example 1.1.5. In shooting at a square target we assume that every shot hits the target and that “every point of the target have an equal probability to be hit.” What is the probability that a shot hits the left lower quadrant of the target? The Laplace definition fails since there are infinitely many points thatwe can hit in the target and infinitely many points in the lower-left quadrant of the target. Dividing infinity by infinity is not helpful. A possible approach is to say that the probability to hit a specific region is proportional to the area of this region. However, this trick works in this example but not sufficiently 6 general to serve as a good definition in many other cases. We will learn later that the correct approach here is to define the prob- ability density at each point of the target. Here is an example that shows that assigning a correct probability den- sity is sometimes not easy and usually should be chosen based on experi- ments. Example 1.1.6 (Bertrand’s paradox). Take a circle of radius 1 and choose at random a chord of it. What is the probability that this chord will be longer than the side of the regular triangle inscribed into the circle? Solution #1: Take a point in the interior of the circle at random and construct the chord whose midpoint is the chosen point. Now we can rotate the triangle so that one of its sides is parallel to the chord. Then it is clear that the chord is longer than the triangle side if and only if the distance from the midpoint of the chord to the center is smaller than the distance from the midpoint of the triangle side to the center. This distance is the radius of the circle inscribed in the triangle which is easy to calculate as 1/2. So the conclusion is that the chord will be longer than the side of the triangle if the midpoint is in the circle with the same center and radius 1/2.
Recommended publications
  • Effective Program Reasoning Using Bayesian Inference
    EFFECTIVE PROGRAM REASONING USING BAYESIAN INFERENCE Sulekha Kulkarni A DISSERTATION in Computer and Information Science Presented to the Faculties of the University of Pennsylvania in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy 2020 Supervisor of Dissertation Mayur Naik, Professor, Computer and Information Science Graduate Group Chairperson Mayur Naik, Professor, Computer and Information Science Dissertation Committee: Rajeev Alur, Zisman Family Professor of Computer and Information Science Val Tannen, Professor of Computer and Information Science Osbert Bastani, Research Assistant Professor of Computer and Information Science Suman Nath, Partner Research Manager, Microsoft Research, Redmond To my father, who set me on this path, to my mother, who leads by example, and to my husband, who is an infinite source of courage. ii Acknowledgments I want to thank my advisor Prof. Mayur Naik for giving me the invaluable opportunity to learn and experiment with different ideas at my own pace. He supported me through the ups and downs of research, and helped me make the Ph.D. a reality. I also want to thank Prof. Rajeev Alur, Prof. Val Tannen, Prof. Osbert Bastani, and Dr. Suman Nath for serving on my dissertation committee and for providing valuable feedback. I am deeply grateful to Prof. Alur and Prof. Tannen for their sound advice and support, and for going out of their way to help me through challenging times. I am also very grateful for Dr. Nath's able and inspiring mentorship during my internship at Microsoft Research, and during the collaboration that followed. Dr. Aditya Nori helped me start my Ph.D.
    [Show full text]
  • 1 Dependent and Independent Events 2 Complementary Events 3 Mutually Exclusive Events 4 Probability of Intersection of Events
    1 Dependent and Independent Events Let A and B be events. We say that A is independent of B if P (AjB) = P (A). That is, the marginal probability of A is the same as the conditional probability of A, given B. This means that the probability of A occurring is not affected by B occurring. It turns out that, in this case, B is independent of A as well. So, we just say that A and B are independent. We say that A depends on B if P (AjB) 6= P (A). That is, the marginal probability of A is not the same as the conditional probability of A, given B. This means that the probability of A occurring is affected by B occurring. It turns out that, in this case, B depends on A as well. So, we just say that A and B are dependent. Consider these events from the card draw: A = drawing a king, B = drawing a spade, C = drawing a face card. Events A and B are independent. If you know that you have drawn a spade, this does not change the likelihood that you have actually drawn a king. Formally, the marginal probability of drawing a king is P (A) = 4=52. The conditional probability that your card is a king, given that it a spade, is P (AjB) = 1=13, which is the same as 4=52. Events A and C are dependent. If you know that you have drawn a face card, it is much more likely that you have actually drawn a king than it would be ordinarily.
    [Show full text]
  • The Open Handbook of Formal Epistemology
    THEOPENHANDBOOKOFFORMALEPISTEMOLOGY Richard Pettigrew &Jonathan Weisberg,Eds. THEOPENHANDBOOKOFFORMAL EPISTEMOLOGY Richard Pettigrew &Jonathan Weisberg,Eds. Published open access by PhilPapers, 2019 All entries copyright © their respective authors and licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. LISTOFCONTRIBUTORS R. A. Briggs Stanford University Michael Caie University of Toronto Kenny Easwaran Texas A&M University Konstantin Genin University of Toronto Franz Huber University of Toronto Jason Konek University of Bristol Hanti Lin University of California, Davis Anna Mahtani London School of Economics Johanna Thoma London School of Economics Michael G. Titelbaum University of Wisconsin, Madison Sylvia Wenmackers Katholieke Universiteit Leuven iii For our teachers Overall, and ultimately, mathematical methods are necessary for philosophical progress. — Hannes Leitgeb There is no mathematical substitute for philosophy. — Saul Kripke PREFACE In formal epistemology, we use mathematical methods to explore the questions of epistemology and rational choice. What can we know? What should we believe and how strongly? How should we act based on our beliefs and values? We begin by modelling phenomena like knowledge, belief, and desire using mathematical machinery, just as a biologist might model the fluc- tuations of a pair of competing populations, or a physicist might model the turbulence of a fluid passing through a small aperture. Then, we ex- plore, discover, and justify the laws governing those phenomena, using the precision that mathematical machinery affords. For example, we might represent a person by the strengths of their beliefs, and we might measure these using real numbers, which we call credences. Having done this, we might ask what the norms are that govern that person when we represent them in that way.
    [Show full text]
  • Lecture 2: Modeling Random Experiments
    Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2021 Lecture 2: Modeling Random Experiments Relevant textbook passages: Pitman [5]: Sections 1.3–1.4., pp. 26–46. Larsen–Marx [4]: Sections 2.2–2.5, pp. 18–66. 2.1 Axioms for probability measures Recall from last time that a random experiment is an experiment that may be conducted under seemingly identical conditions, yet give different results. Coin tossing is everyone’s go-to example of a random experiment. The way we model random experiments is through the use of probabilities. We start with the sample space Ω, the set of possible outcomes of the experiment, and consider events, which are subsets E of the sample space. (We let F denote the collection of events.) 2.1.1 Definition A probability measure P or probability distribution attaches to each event E a number between 0 and 1 (inclusive) so as to obey the following axioms of probability: Normalization: P (?) = 0; and P (Ω) = 1. Nonnegativity: For each event E, we have P (E) > 0. Additivity: If EF = ?, then P (∪ F ) = P (E) + P (F ). Note that while the domain of P is technically F, the set of events, that is P : F → [0, 1], we may also refer to P as a probability (measure) on Ω, the set of realizations. 2.1.2 Remark To reduce the visual clutter created by layers of delimiters in our notation, we may omit some of them simply write something like P (f(ω) = 1) orP {ω ∈ Ω: f(ω) = 1} instead of P {ω ∈ Ω: f(ω) = 1} and we may write P (ω) instead of P {ω} .
    [Show full text]
  • Probability and Counting Rules
    blu03683_ch04.qxd 09/12/2005 12:45 PM Page 171 C HAPTER 44 Probability and Counting Rules Objectives Outline After completing this chapter, you should be able to 4–1 Introduction 1 Determine sample spaces and find the probability of an event, using classical 4–2 Sample Spaces and Probability probability or empirical probability. 4–3 The Addition Rules for Probability 2 Find the probability of compound events, using the addition rules. 4–4 The Multiplication Rules and Conditional 3 Find the probability of compound events, Probability using the multiplication rules. 4–5 Counting Rules 4 Find the conditional probability of an event. 5 Find the total number of outcomes in a 4–6 Probability and Counting Rules sequence of events, using the fundamental counting rule. 4–7 Summary 6 Find the number of ways that r objects can be selected from n objects, using the permutation rule. 7 Find the number of ways that r objects can be selected from n objects without regard to order, using the combination rule. 8 Find the probability of an event, using the counting rules. 4–1 blu03683_ch04.qxd 09/12/2005 12:45 PM Page 172 172 Chapter 4 Probability and Counting Rules Statistics Would You Bet Your Life? Today Humans not only bet money when they gamble, but also bet their lives by engaging in unhealthy activities such as smoking, drinking, using drugs, and exceeding the speed limit when driving. Many people don’t care about the risks involved in these activities since they do not understand the concepts of probability.
    [Show full text]
  • 1 — a Single Random Variable
    1 | A SINGLE RANDOM VARIABLE Questions involving probability abound in Computer Science: What is the probability of the PWF world falling over next week? • What is the probability of one packet colliding with another in a network? • What is the probability of an undergraduate not turning up for a lecture? • When addressing such questions there are often added complications: the question may be ill posed or the answer may vary with time. Which undergraduate? What lecture? Is the probability of turning up different on Saturdays? Let's start with something which appears easy to reason about: : : Introduction | Throwing a die Consider an experiment or trial which consists of throwing a mathematically ideal die. Such a die is often called a fair die or an unbiased die. Common sense suggests that: The outcome of a single throw cannot be predicted. • The outcome will necessarily be a random integer in the range 1 to 6. • The six possible outcomes are equiprobable, each having a probability of 1 . • 6 Without further qualification, serious probabilists would regard this collection of assertions, especially the second, as almost meaningless. Just what is a random integer? Giving proper mathematical rigour to the foundations of probability theory is quite a taxing task. To illustrate the difficulty, consider probability in a frequency sense. Thus a probability 1 of 6 means that, over a long run, one expects to throw a 5 (say) on one-sixth of the occasions that the die is thrown. If the actual proportion of 5s after n throws is p5(n) it would be nice to say: 1 lim p5(n) = n !1 6 Unfortunately this is utterly bogus mathematics! This is simply not a proper use of the idea of a limit.
    [Show full text]
  • Probabilities, Random Variables and Distributions A
    Probabilities, Random Variables and Distributions A Contents A.1 EventsandProbabilities................................ 318 A.1.1 Conditional Probabilities and Independence . ............. 318 A.1.2 Bayes’Theorem............................... 319 A.2 Random Variables . ................................. 319 A.2.1 Discrete Random Variables ......................... 319 A.2.2 Continuous Random Variables ....................... 320 A.2.3 TheChangeofVariablesFormula...................... 321 A.2.4 MultivariateNormalDistributions..................... 323 A.3 Expectation,VarianceandCovariance........................ 324 A.3.1 Expectation................................. 324 A.3.2 Variance................................... 325 A.3.3 Moments................................... 325 A.3.4 Conditional Expectation and Variance ................... 325 A.3.5 Covariance.................................. 326 A.3.6 Correlation.................................. 327 A.3.7 Jensen’sInequality............................. 328 A.3.8 Kullback–LeiblerDiscrepancyandInformationInequality......... 329 A.4 Convergence of Random Variables . 329 A.4.1 Modes of Convergence . 329 A.4.2 Continuous Mapping and Slutsky’s Theorem . 330 A.4.3 LawofLargeNumbers........................... 330 A.4.4 CentralLimitTheorem........................... 331 A.4.5 DeltaMethod................................ 331 A.5 ProbabilityDistributions............................... 332 A.5.1 UnivariateDiscreteDistributions...................... 333 A.5.2 Univariate Continuous Distributions . 335
    [Show full text]
  • (Introduction to Probability at an Advanced Level) - All Lecture Notes
    Fall 2018 Statistics 201A (Introduction to Probability at an advanced level) - All Lecture Notes Aditya Guntuboyina August 15, 2020 Contents 0.1 Sample spaces, Events, Probability.................................5 0.2 Conditional Probability and Independence.............................6 0.3 Random Variables..........................................7 1 Random Variables, Expectation and Variance8 1.1 Expectations of Random Variables.................................9 1.2 Variance................................................ 10 2 Independence of Random Variables 11 3 Common Distributions 11 3.1 Ber(p) Distribution......................................... 11 3.2 Bin(n; p) Distribution........................................ 11 3.3 Poisson Distribution......................................... 12 4 Covariance, Correlation and Regression 14 5 Correlation and Regression 16 6 Back to Common Distributions 16 6.1 Geometric Distribution........................................ 16 6.2 Negative Binomial Distribution................................... 17 7 Continuous Distributions 17 7.1 Normal or Gaussian Distribution.................................. 17 1 7.2 Uniform Distribution......................................... 18 7.3 The Exponential Density...................................... 18 7.4 The Gamma Density......................................... 18 8 Variable Transformations 19 9 Distribution Functions and the Quantile Transform 20 10 Joint Densities 22 11 Joint Densities under Transformations 23 11.1 Detour to Convolutions......................................
    [Show full text]
  • Why Retail Therapy Works: It Is Choice, Not Acquisition, That
    ASSOCIATION FOR CONSUMER RESEARCH Labovitz School of Business & Economics, University of Minnesota Duluth, 11 E. Superior Street, Suite 210, Duluth, MN 55802 Why Retail Therapy Works: It Is Choice, Not Acquisition, That Primarily Alleviates Sadness Beatriz Pereira, University of Michigan, USA Scott Rick, University of Michigan, USA Can shopping be used strategically as an emotion regulation tool? Although prior work demonstrates that sadness encourages spending, it is unclear whether and why shopping actually alleviates sadness. Our work suggests that shopping can heal, but that it is the act of choosing (e.g., between money and products), rather than the act of acquiring (e.g., simply being endowed with money or products), that primarily alleviates sadness. Two experiments that induced sadness and then manipulated whether participants made monetarily consequential choices support our conclusions. [to cite]: Beatriz Pereira and Scott Rick (2011) ,"Why Retail Therapy Works: It Is Choice, Not Acquisition, That Primarily Alleviates Sadness", in NA - Advances in Consumer Research Volume 39, eds. Rohini Ahluwalia, Tanya L. Chartrand, and Rebecca K. Ratner, Duluth, MN : Association for Consumer Research, Pages: 732-733. [url]: http://www.acrwebsite.org/volumes/1009733/volumes/v39/NA-39 [copyright notice]: This work is copyrighted by The Association for Consumer Research. For permission to copy or use this work in whole or in part, please contact the Copyright Clearance Center at http://www.copyright.com/. 732 / Working Papers SIGNIFICANCE AND Implications OF THE RESEARCH In this study, we examine how people’s judgment on the probability of a conjunctive event influences their subsequent inference (e.g., after successfully getting five papers accepted what is the probability of getting tenure?).
    [Show full text]
  • Probability with Engineering Applications ECE 313 Course Notes
    Probability with Engineering Applications ECE 313 Course Notes Bruce Hajek Department of Electrical and Computer Engineering University of Illinois at Urbana-Champaign January 2017 c 2017 by Bruce Hajek All rights reserved. Permission is hereby given to freely print and circulate copies of these notes so long as the notes are left intact and not reproduced for commercial purposes. Email to [email protected], pointing out errors or hard to understand passages or providing comments, is welcome. Contents 1 Foundations 3 1.1 Embracing uncertainty . .3 1.2 Axioms of probability . .6 1.3 Calculating the size of various sets . 10 1.4 Probability experiments with equally likely outcomes . 13 1.5 Sample spaces with infinite cardinality . 15 1.6 Short Answer Questions . 20 1.7 Problems . 21 2 Discrete-type random variables 25 2.1 Random variables and probability mass functions . 25 2.2 The mean and variance of a random variable . 27 2.3 Conditional probabilities . 32 2.4 Independence and the binomial distribution . 34 2.4.1 Mutually independent events . 34 2.4.2 Independent random variables (of discrete-type) . 36 2.4.3 Bernoulli distribution . 37 2.4.4 Binomial distribution . 38 2.5 Geometric distribution . 41 2.6 Bernoulli process and the negative binomial distribution . 43 2.7 The Poisson distribution{a limit of binomial distributions . 45 2.8 Maximum likelihood parameter estimation . 47 2.9 Markov and Chebychev inequalities and confidence intervals . 50 2.10 The law of total probability, and Bayes formula . 53 2.11 Binary hypothesis testing with discrete-type observations .
    [Show full text]
  • Chapter 2: Probability
    16 Chapter 2: Probability The aim of this chapter is to revise the basic rules of probability. By the end of this chapter, you should be comfortable with: • conditional probability, and what you can and can’t do with conditional expressions; • the Partition Theorem and Bayes’ Theorem; • First-Step Analysis for finding the probability that a process reaches some state, by conditioning on the outcome of the first step; • calculating probabilities for continuous and discrete random variables. 2.1 Sample spaces and events Definition: A sample space, Ω, is a set of possible outcomes of a random experiment. Definition: An event, A, is a subset of the sample space. This means that event A is simply a collection of outcomes. Example: Random experiment: Pick a person in this class at random. Sample space: Ω= {all people in class} Event A: A = {all males in class}. Definition: Event A occurs if the outcome of the random experiment is a member of the set A. In the example above, event A occurs if the person we pick is male. 17 2.2 Probability Reference List The following properties hold for all events A, B. • P(∅)=0. • 0 ≤ P(A) ≤ 1. • Complement: P(A)=1 − P(A). • Probability of a union: P(A ∪ B)= P(A)+ P(B) − P(A ∩ B). For three events A, B, C: P(A∪B∪C)= P(A)+P(B)+P(C)−P(A∩B)−P(A∩C)−P(B∩C)+P(A∩B∩C) . If A and B are mutually exclusive, then P(A ∪ B)= P(A)+ P(B).
    [Show full text]
  • Probabilities and Expectations
    Probabilities and Expectations Ashique Rupam Mahmood September 9, 2015 Probabilities tell us about the likelihood of an event in numbers. If an event is certain to occur, such as sunrise, probability of that event is said to be 1. Pr(sunrise) = 1. If an event will certainly not occur, then its probability is 0. So, probability maps events to a number in [0; 1]. How do you specify an event? In the discussions of probabilities, events are technically described as a set. At this point it is important to go through some basic concepts of sets and maybe also functions. Sets A set is a collection of distinct objects. For example, if we toss a coin once, the set of all possible distinct outcomes will be S = head; tail , where head denotes a head and the f g tail denotes a tail. All sets we consider here are finite. An element of a set is denoted as head S.A subset of a set is denoted as head S. 2 f g ⊂ What are the possible subsets of S? These are: head ; tail ;S = head; tail , and f g f g f g φ = . So, note that a set is a subset of itself: S S. Also note that, an empty set fg ⊂ (a collection of nothing) is a subset of any set: φ S.A union of two sets A and B is ⊂ comprised of all the elements of both sets and denoted as A B. An intersection of two [ sets A and B is comprised of only the common elements of both sets and denoted as A B.
    [Show full text]