Lecture 10 Conditional Expectation

Total Page:16

File Type:pdf, Size:1020Kb

Lecture 10 Conditional Expectation Lecture 10:Conditional Expectation 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 10 Conditional Expectation The definition and existence of conditional expectation For events A, B with P[B] > 0, we recall the familiar object P[ j ] = P[A\B] A B P[B] . We say that P[AjB] the conditional probability of A, given B. It is important to note that the condition P[B] > 0 is crucial. When X and Y are random variables defined on the same probability space, we often want to give a meaning to the expression P[X 2 AjY = y], even though it is usually the case that P[Y = y] = 0. When the random vector (X, Y) admits a joint density fX,Y(x, y), and fY(y) > 0, the concept of conditional density fXjY=y(x) = fX,Y(x, y)/ fY(y) is introduced and R the quantity P[X 2 AjY = y] is given meaning via A fXjY=y(x, y) dx. While this procedure works well in the restrictive case of absolutely continuous random vectors, we will see how it is encompassed by a general concept of a conditional expectation. Since probability is simply an expectation of an indicator, and expectations are linear, it will be easier to work with expectations and no generality will be lost. Two main conceptual leaps here are: 1) we condition with respect to a s-algebra, and 2) we view the conditional expectation itself as a random variable. Before we illustrate the concept in discrete time, here is the definition. Definition 10.1. Let G be a sub-s-algebra of F, and let X 2 L1 be a random variable. We say that the random variable x is (a version of) the conditional expectation of X with respect to G - and denote it by E[XjG] - if 1. x 2 L1. 2. x is G-measurable, 3. E[x1A] = E[X1A], for all A 2 G. Last Updated: January 24, 2015 Lecture 10:Conditional Expectation 2 of 17 Example 10.2. Suppose that (W, F, P) is a probability space where X W = fa, b, c, d, e, f g, F = 2W and P is uniform. Let X, Y and Z be E[X (Y )] random variables given by (in the obvious notation) ! a b c d e f X ∼ , 1 3 3 5 5 7 ! ! a b c d e f a b c d e f Y ∼ and Z ∼ 2 2 1 1 7 7 3 3 3 3 2 2 We would like to think about E[XjG] as the average of X(w) over all w which are consistent with the our current information (which is G). For G = (Y) G example, if s , then the information contained in is exactly X the information about the exact value of Y. Knowledge of the fact that [X ( )] Y = y does not necessarily reveal the “true” w, but certainly rules out E all those w for which Y(w) 6= y. In our specific case, if we know that Y = 2, then w = a or w = b, 1 1 and the expected value of X, given that Y = 2, is 2 X(a) + 2 X(b) = 2. Similarly, this average equals 4 for Y = 1, and 6 for Y = 7. Let us show that the random variable x defined by this average, i.e., ! a b c d e f x ∼ , 2 2 4 4 6 6 satisfies the definition of E[Xjs(Y)], as given above. The integrability is not an issue (we are on a finite probability space), and it is clear that x is measurable with respect to s(Y). Indeed, the atoms of s(Y) are fa, bg, fc, dg and fe, f g, and x is constant over each one of those. Finally, we need to check that E[x1A] = E[X1A], for all A 2 s(Y), which for an atom A translates into ( ) = 1 E[ ] = ( 0)P[f 0gj ] 2 x w P[A] X1A ∑ X w w A , for all w A. w02A The moral of the story is that when A is an atom, part 3. of Defini- tion 10.1 translates into a requirement that x be constant on A with value equal to the expectation of X over A with respect to the condi- tional probability P[·jA]. In the general case, when there are no atoms, 3. still makes sense and conveys the same message. Btw, since the atoms of s(Z) are fa, b, c, dg and fe, f g, it is clear that 8 <3, w 2 fa, b, c, dg, E[Xjs(Z)](w) = :6, w 2 fe, f g. Last Updated: January 24, 2015 Lecture 10:Conditional Expectation 3 of 17 Look at the illustrations above and convince yourself that E[E[Xjs(Y)]js(Z)] = E[Xjs(Z)]. A general result along the same lines - called the tower property of con- ditional expectation - will be stated and proved below. Our first task is to prove that conditional expectations always exist. When W is finite (as explained above) or countable, we can always construct them by averaging over atoms. In the general case, a different argument is needed. In fact, here are two: Proposition 10.3. Let G be a sub-s-algebra G of F. Then 1. there exists a conditional expectation E[XjG] for any X 2 L1, and 2. any two conditional expectations of X 2 L1 are equal P-a.s. Proof. (Uniqueness): Suppose that x and x0 both satisfy 1., 2. and 3. of Definition 10.1. Then 0 E[x1A] = E[x 1A], for all A 2 G. 0 1 For An = fx − x ≥ n g, we have An 2 G and so E[ ] = E[ 0 ] ≥ E[( + 1 ) ] = E[ ] + 1 P[ ] x1An x 1An x n 1An x1An n An . 0 Consequently, P[An] = 0, for all n 2 N, so that P[x > x] = 0. By a symmetric argument, we also have P[x0 < x] = 0. (Existence): By linearity, it will be enough to prove that the condi- 1 tional expectation exists for X 2 L+. 1. A Radon-Nikodym argument. Suppose, first, that X ≥ 0 and E[X] = 1, as the general case follows by additivity and scaling. Then the prescription Q[A] = E[X1A], defines a probability measure on (W, F), which is absolutely continu- ous with respect to P. Let QG be the restriction of Q to G; it is trivially absolutely continuous with respect to the restriction PG of P to G. The Radon-Nikodym theorem - applied to the measure space (W, G, PG ) and the measure QG PG - guarantees the existence of the Radon- Nikodym derivative dQG x = 2 L1 (W, G, PG ). dPG + For A 2 G, we thus have G PG E[X1A] = Q[A] = Q [A] = E [x1A] = E[x1A]. Last Updated: January 24, 2015 Lecture 10:Conditional Expectation 4 of 17 where the last equality follows from the fact that x1A is G-measurable. Therefore, x is (a version of) the conditional expectation E[XjG]. 1. An L2-argument. Suppose, first, that X 2 L2. Let H be the family of all G-measurable elements in L2. Let H¯ denote the closure of H in the topology induced by L2-convergence. Being a closed and convex (why?) subset of L2, H¯ satisfies all the conditions of Problem ?? so that there exists x 2 H¯ at the minimal L2-distance from X (when X 2 H¯ , we take x = X). The same problem states that x has the following property: E[(h − x)(X − x)] ≥ 0 for all h 2 H¯ , and, since H¯ is a linear space, we have E[(h − x)(X − x)] = 0, for all h 2 H¯ . It remains to pick h of the form h = x + 1A 2 H¯ , A 2 G, to conclude that E[X1A] = E[x1A], for all A 2 G. Our next step is to show that x is G-measurable (after a modifi- cation on a null set, perhaps). Since x 2 H¯ , there exists a sequence 2 a.s. fxngn2N such that xn ! x in L . By Corollary ??, xnk ! x, for 0 some subsequence fxnk gk2N of fxngn2N. Set x = lim infk2N xnk 2 0 ˆ 0 ˆ ˆ L ([−¥, ¥], G) and x = x 1fjx0j<¥g, so that x = x, a.s., and x is G- measurable. 2 We still need to remove the restriction X 2 L+. We start with 1 ¥ 2 a general X 2 L+ and define Xn = min(X, n) 2 L+ ⊆ L+. Let xn = E[XnjG], and note that E[xn+11A] = E[Xn+11A] ≥ E[Xn1A] = E[xn1A]. It follows (just like in the proof of uniqueness above) that xn ≤ xn+1, a.s. We define x = supn xn, so that xn % x, a.s. Then, for A 2 G, the monotone-convergence theorem implies that E[X1A] = lim E[Xn1A] = lim E[xn1A] = E[x1A], n n 1 and it is easy to check that x1fx<¥g 2 L (G) is a version of E[XjG]. Remark 10.4. There is no canonical way to choose “the version” of the conditional expectation. We follow the convention started with Radon- Nikodym derivatives, and interpret a statement such at x ≤ E[XjG], a.s., to mean that x ≤ x0, a.s., for any version x0 of the conditional expectation of X with respect to G. If we use the symbol L1 to denote the set of all a.s.-equivalence classes of random variables in L1, we can write: E[·jG] : L1(F) ! L1(G), Last Updated: January 24, 2015 Lecture 10:Conditional Expectation 5 of 17 but L1(G) cannot be replaced by L1(G) in a natural way.
Recommended publications
  • Lecture 22: Bivariate Normal Distribution Distribution
    6.5 Conditional Distributions General Bivariate Normal Let Z1; Z2 ∼ N (0; 1), which we will use to build a general bivariate normal Lecture 22: Bivariate Normal Distribution distribution. 1 1 2 2 f (z1; z2) = exp − (z1 + z2 ) Statistics 104 2π 2 We want to transform these unit normal distributions to have the follow Colin Rundel arbitrary parameters: µX ; µY ; σX ; σY ; ρ April 11, 2012 X = σX Z1 + µX p 2 Y = σY [ρZ1 + 1 − ρ Z2] + µY Statistics 104 (Colin Rundel) Lecture 22 April 11, 2012 1 / 22 6.5 Conditional Distributions 6.5 Conditional Distributions General Bivariate Normal - Marginals General Bivariate Normal - Cov/Corr First, lets examine the marginal distributions of X and Y , Second, we can find Cov(X ; Y ) and ρ(X ; Y ) Cov(X ; Y ) = E [(X − E(X ))(Y − E(Y ))] X = σX Z1 + µX h p i = E (σ Z + µ − µ )(σ [ρZ + 1 − ρ2Z ] + µ − µ ) = σX N (0; 1) + µX X 1 X X Y 1 2 Y Y 2 h p 2 i = N (µX ; σX ) = E (σX Z1)(σY [ρZ1 + 1 − ρ Z2]) h 2 p 2 i = σX σY E ρZ1 + 1 − ρ Z1Z2 p 2 2 Y = σY [ρZ1 + 1 − ρ Z2] + µY = σX σY ρE[Z1 ] p 2 = σX σY ρ = σY [ρN (0; 1) + 1 − ρ N (0; 1)] + µY = σ [N (0; ρ2) + N (0; 1 − ρ2)] + µ Y Y Cov(X ; Y ) ρ(X ; Y ) = = ρ = σY N (0; 1) + µY σX σY 2 = N (µY ; σY ) Statistics 104 (Colin Rundel) Lecture 22 April 11, 2012 2 / 22 Statistics 104 (Colin Rundel) Lecture 22 April 11, 2012 3 / 22 6.5 Conditional Distributions 6.5 Conditional Distributions General Bivariate Normal - RNG Multivariate Change of Variables Consequently, if we want to generate a Bivariate Normal random variable Let X1;:::; Xn have a continuous joint distribution with pdf f defined of S.
    [Show full text]
  • Bayes and the Law
    Bayes and the Law Norman Fenton, Martin Neil and Daniel Berger [email protected] January 2016 This is a pre-publication version of the following article: Fenton N.E, Neil M, Berger D, “Bayes and the Law”, Annual Review of Statistics and Its Application, Volume 3, 2016, doi: 10.1146/annurev-statistics-041715-033428 Posted with permission from the Annual Review of Statistics and Its Application, Volume 3 (c) 2016 by Annual Reviews, http://www.annualreviews.org. Abstract Although the last forty years has seen considerable growth in the use of statistics in legal proceedings, it is primarily classical statistical methods rather than Bayesian methods that have been used. Yet the Bayesian approach avoids many of the problems of classical statistics and is also well suited to a broader range of problems. This paper reviews the potential and actual use of Bayes in the law and explains the main reasons for its lack of impact on legal practice. These include misconceptions by the legal community about Bayes’ theorem, over-reliance on the use of the likelihood ratio and the lack of adoption of modern computational methods. We argue that Bayesian Networks (BNs), which automatically produce the necessary Bayesian calculations, provide an opportunity to address most concerns about using Bayes in the law. Keywords: Bayes, Bayesian networks, statistics in court, legal arguments 1 1 Introduction The use of statistics in legal proceedings (both criminal and civil) has a long, but not terribly well distinguished, history that has been very well documented in (Finkelstein, 2009; Gastwirth, 2000; Kadane, 2008; Koehler, 1992; Vosk and Emery, 2014).
    [Show full text]
  • Estimating the Accuracy of Jury Verdicts
    Institute for Policy Research Northwestern University Working Paper Series WP-06-05 Estimating the Accuracy of Jury Verdicts Bruce D. Spencer Faculty Fellow, Institute for Policy Research Professor of Statistics Northwestern University Version date: April 17, 2006; rev. May 4, 2007 Forthcoming in Journal of Empirical Legal Studies 2040 Sheridan Rd. ! Evanston, IL 60208-4100 ! Tel: 847-491-3395 Fax: 847-491-9916 www.northwestern.edu/ipr, ! [email protected] Abstract Average accuracy of jury verdicts for a set of cases can be studied empirically and systematically even when the correct verdict cannot be known. The key is to obtain a second rating of the verdict, for example the judge’s, as in the recent study of criminal cases in the U.S. by the National Center for State Courts (NCSC). That study, like the famous Kalven-Zeisel study, showed only modest judge-jury agreement. Simple estimates of jury accuracy can be developed from the judge-jury agreement rate; the judge’s verdict is not taken as the gold standard. Although the estimates of accuracy are subject to error, under plausible conditions they tend to overestimate the average accuracy of jury verdicts. The jury verdict was estimated to be accurate in no more than 87% of the NCSC cases (which, however, should not be regarded as a representative sample with respect to jury accuracy). More refined estimates, including false conviction and false acquittal rates, are developed with models using stronger assumptions. For example, the conditional probability that the jury incorrectly convicts given that the defendant truly was not guilty (a “type I error”) was estimated at 0.25, with an estimated standard error (s.e.) of 0.07, the conditional probability that a jury incorrectly acquits given that the defendant truly was guilty (“type II error”) was estimated at 0.14 (s.e.
    [Show full text]
  • General Probability, II: Independence and Conditional Proba- Bility
    Math 408, Actuarial Statistics I A.J. Hildebrand General Probability, II: Independence and conditional proba- bility Definitions and properties 1. Independence: A and B are called independent if they satisfy the product formula P (A ∩ B) = P (A)P (B). 2. Conditional probability: The conditional probability of A given B is denoted by P (A|B) and defined by the formula P (A ∩ B) P (A|B) = , P (B) provided P (B) > 0. (If P (B) = 0, the conditional probability is not defined.) 3. Independence of complements: If A and B are independent, then so are A and B0, A0 and B, and A0 and B0. 4. Connection between independence and conditional probability: If the con- ditional probability P (A|B) is equal to the ordinary (“unconditional”) probability P (A), then A and B are independent. Conversely, if A and B are independent, then P (A|B) = P (A) (assuming P (B) > 0). 5. Complement rule for conditional probabilities: P (A0|B) = 1 − P (A|B). That is, with respect to the first argument, A, the conditional probability P (A|B) satisfies the ordinary complement rule. 6. Multiplication rule: P (A ∩ B) = P (A|B)P (B) Some special cases • If P (A) = 0 or P (B) = 0 then A and B are independent. The same holds when P (A) = 1 or P (B) = 1. • If B = A or B = A0, A and B are not independent except in the above trivial case when P (A) or P (B) is 0 or 1. In other words, an event A which has probability strictly between 0 and 1 is not independent of itself or of its complement.
    [Show full text]
  • Laws of Total Expectation and Total Variance
    Laws of Total Expectation and Total Variance Definition of conditional density. Assume and arbitrary random variable X with density fX . Take an event A with P (A) > 0. Then the conditional density fXjA is defined as follows: 8 < f(x) 2 P (A) x A fXjA(x) = : 0 x2 = A Note that the support of fXjA is supported only in A. Definition of conditional expectation conditioned on an event. Z Z 1 E(h(X)jA) = h(x) fXjA(x) dx = h(x) fX (x) dx A P (A) A Example. For the random variable X with density function 8 < 1 3 64 x 0 < x < 4 f(x) = : 0 otherwise calculate E(X2 j X ≥ 1). Solution. Step 1. Z h i 4 1 1 1 x=4 255 P (X ≥ 1) = x3 dx = x4 = 1 64 256 4 x=1 256 Step 2. Z Z ( ) 1 256 4 1 E(X2 j X ≥ 1) = x2 f(x) dx = x2 x3 dx P (X ≥ 1) fx≥1g 255 1 64 Z ( ) ( )( ) h i 256 4 1 256 1 1 x=4 8192 = x5 dx = x6 = 255 1 64 255 64 6 x=1 765 Definition of conditional variance conditioned on an event. Var(XjA) = E(X2jA) − E(XjA)2 1 Example. For the previous example , calculate the conditional variance Var(XjX ≥ 1) Solution. We already calculated E(X2 j X ≥ 1). We only need to calculate E(X j X ≥ 1). Z Z Z ( ) 1 256 4 1 E(X j X ≥ 1) = x f(x) dx = x x3 dx P (X ≥ 1) fx≥1g 255 1 64 Z ( ) ( )( ) h i 256 4 1 256 1 1 x=4 4096 = x4 dx = x5 = 255 1 64 255 64 5 x=1 1275 Finally: ( ) 8192 4096 2 630784 Var(XjX ≥ 1) = E(X2jX ≥ 1) − E(XjX ≥ 1)2 = − = 765 1275 1625625 Definition of conditional expectation conditioned on a random variable.
    [Show full text]
  • Randomized Algorithms
    Chapter 9 Randomized Algorithms randomized The theme of this chapter is madendrizo algorithms. These are algorithms that make use of randomness in their computation. You might know of quicksort, which is efficient on average when it uses a random pivot, but can be bad for any pivot that is selected without randomness. Analyzing randomized algorithms can be difficult, so you might wonder why randomiza- tion in algorithms is so important and worth the extra effort. Well it turns out that for certain problems randomized algorithms are simpler or faster than algorithms that do not use random- ness. The problem of primality testing (PT), which is to determine if an integer is prime, is a good example. In the late 70s Miller and Rabin developed a famous and simple random- ized algorithm for the problem that only requires polynomial work. For over 20 years it was not known whether the problem could be solved in polynomial work without randomization. Eventually a polynomial time algorithm was developed, but it is much more complicated and computationally more costly than the randomized version. Hence in practice everyone still uses the randomized version. There are many other problems in which a randomized solution is simpler or cheaper than the best non-randomized solution. In this chapter, after covering the prerequisite background, we will consider some such problems. The first we will consider is the following simple prob- lem: Question: How many comparisons do we need to find the top two largest numbers in a sequence of n distinct numbers? Without the help of randomization, there is a trivial algorithm for finding the top two largest numbers in a sequence that requires about 2n − 3 comparisons.
    [Show full text]
  • Propensities and Probabilities
    ARTICLE IN PRESS Studies in History and Philosophy of Modern Physics 38 (2007) 593–625 www.elsevier.com/locate/shpsb Propensities and probabilities Nuel Belnap 1028-A Cathedral of Learning, University of Pittsburgh, Pittsburgh, PA 15260, USA Received 19 May 2006; accepted 6 September 2006 Abstract Popper’s introduction of ‘‘propensity’’ was intended to provide a solid conceptual foundation for objective single-case probabilities. By considering the partly opposed contributions of Humphreys and Miller and Salmon, it is argued that when properly understood, propensities can in fact be understood as objective single-case causal probabilities of transitions between concrete events. The chief claim is that propensities are well-explicated by describing how they fit into the existing formal theory of branching space-times, which is simultaneously indeterministic and causal. Several problematic examples, some commonsense and some quantum-mechanical, are used to make clear the advantages of invoking branching space-times theory in coming to understand propensities. r 2007 Elsevier Ltd. All rights reserved. Keywords: Propensities; Probabilities; Space-times; Originating causes; Indeterminism; Branching histories 1. Introduction You are flipping a fair coin fairly. You ascribe a probability to a single case by asserting The probability that heads will occur on this very next flip is about 50%. ð1Þ The rough idea of a single-case probability seems clear enough when one is told that the contrast is with either generalizations or frequencies attributed to populations asserted while you are flipping a fair coin fairly, such as In the long run; the probability of heads occurring among flips is about 50%. ð2Þ E-mail address: [email protected] 1355-2198/$ - see front matter r 2007 Elsevier Ltd.
    [Show full text]
  • MA3K0 - High-Dimensional Probability
    MA3K0 - High-Dimensional Probability Lecture Notes Stefan Adams i 2020, update 06.05.2020 Notes are in final version - proofreading not completed yet! Typos and errors will be updated on a regular basis Contents 1 Prelimaries on Probability Theory 1 1.1 Random variables . .1 1.2 Classical Inequalities . .4 1.3 Limit Theorems . .6 2 Concentration inequalities for independent random variables 8 2.1 Why concentration inequalities . .8 2.2 Hoeffding’s Inequality . 10 2.3 Chernoff’s Inequality . 13 2.4 Sub-Gaussian random variables . 15 2.5 Sub-Exponential random variables . 22 3 Random vectors in High Dimensions 27 3.1 Concentration of the Euclidean norm . 27 3.2 Covariance matrices and Principal Component Analysis (PCA) . 30 3.3 Examples of High-Dimensional distributions . 32 3.4 Sub-Gaussian random variables in higher dimensions . 34 3.5 Application: Grothendieck’s inequality . 36 4 Random Matrices 36 4.1 Geometrics concepts . 36 4.2 Concentration of the operator norm of random matrices . 38 4.3 Application: Community Detection in Networks . 41 4.4 Application: Covariance Estimation and Clustering . 41 5 Concentration of measure - general case 43 5.1 Concentration by entropic techniques . 43 5.2 Concentration via Isoperimetric Inequalities . 50 5.3 Some matrix calculus and covariance estimation . 54 6 Basic tools in high-dimensional probability 57 6.1 Decoupling . 57 6.2 Concentration for Anisotropic random vectors . 62 6.3 Symmetrisation . 63 7 Random Processes 64 ii Preface Introduction We discuss an elegant argument that showcases the
    [Show full text]
  • Conditional Probability and Bayes Theorem A
    Conditional probability And Bayes theorem A. Zaikin 2.1 Conditional probability 1 Conditional probablity Given events E and F ,oftenweareinterestedinstatementslike if even E has occurred, then the probability of F is ... Some examples: • Roll two dice: what is the probability that the sum of faces is 6 given that the first face is 4? • Gene expressions: What is the probability that gene A is switched off (e.g. down-regulated) given that gene B is also switched off? A. Zaikin 2.2 Conditional probability 2 This conditional probability can be derived following a similar construction: • Repeat the experiment N times. • Count the number of times event E occurs, N(E),andthenumberoftimesboth E and F occur jointly, N(E ∩ F ).HenceN(E) ≤ N • The proportion of times that F occurs in this reduced space is N(E ∩ F ) N(E) since E occurs at each one of them. • Now note that the ratio above can be re-written as the ratio between two (unconditional) probabilities N(E ∩ F ) N(E ∩ F )/N = N(E) N(E)/N • Then the probability of F ,giventhatE has occurred should be defined as P (E ∩ F ) P (E) A. Zaikin 2.3 Conditional probability: definition The definition of Conditional Probability The conditional probability of an event F ,giventhataneventE has occurred, is defined as P (E ∩ F ) P (F |E)= P (E) and is defined only if P (E) > 0. Note that, if E has occurred, then • F |E is a point in the set P (E ∩ F ) • E is the new sample space it can be proved that the function P (·|·) defyning a conditional probability also satisfies the three probability axioms.
    [Show full text]
  • Multicollinearity
    Chapter 10 Multicollinearity 10.1 The Nature of Multicollinearity 10.1.1 Extreme Collinearity The standard OLS assumption that ( xi1; xi2; : : : ; xik ) not be linearly related means that for any ( c1; c2; : : : ; ck ) xik =6 c1xi1 + c2xi2 + ··· + ck 1xi;k 1 (10.1) − − for some i. If the assumption is violated, then we can find ( c1; c2; : : : ; ck 1 ) such that − xik = c1xi1 + c2xi2 + ··· + ck 1xi;k 1 (10.2) − − for all i. Define x12 ··· x1k xk1 c1 x22 ··· x2k xk2 c2 X1 = 0 . 1 ; xk = 0 . 1 ; and c = 0 . 1 : . B C B C B C B xn2 ··· xnk C B xkn C B ck 1 C B C B C B − C @ A @ A @ A Then extreme collinearity can be represented as xk = X1c: (10.3) We have represented extreme collinearity in terms of the last explanatory vari- able. Since we can always re-order the variables this choice is without loss of generality and the analysis could be applied to any non-constant variable by moving it to the last column. 10.1.2 Near Extreme Collinearity Of course, it is rare, in practice, that an exact linear relationship holds. Instead, we have xik = c1xi1 + c2xi2 + ··· + ck 1xi;k 1 + vi (10.4) − − 133 134 CHAPTER 10. MULTICOLLINEARITY or, more compactly, xk = X1c + v; (10.5) where the v's are small relative to the x's. If we think of the v's as random variables they will have small variance (and zero mean if X includes a column of ones). A convenient way to algebraically express the degree of collinearity is the sample correlation between xik and wi = c1xi1 +c2xi2 +···+ck 1xi;k 1, namely − − cov( xik; wi ) cov( wi + vi; wi ) rx;w = = (10.6) var(xi;k)var(wi) var(wi + vi)var(wi) d d Clearly, as the variancep of vi grows small, thisp value will go to unity.
    [Show full text]
  • CONDITIONAL EXPECTATION Definition 1. Let (Ω,F,P)
    CONDITIONAL EXPECTATION 1. CONDITIONAL EXPECTATION: L2 THEORY ¡ Definition 1. Let (­,F ,P) be a probability space and let G be a σ algebra contained in F . For ¡ any real random variable X L2(­,F ,P), define E(X G ) to be the orthogonal projection of X 2 j onto the closed subspace L2(­,G ,P). This definition may seem a bit strange at first, as it seems not to have any connection with the naive definition of conditional probability that you may have learned in elementary prob- ability. However, there is a compelling rationale for Definition 1: the orthogonal projection E(X G ) minimizes the expected squared difference E(X Y )2 among all random variables Y j ¡ 2 L2(­,G ,P), so in a sense it is the best predictor of X based on the information in G . It may be helpful to consider the special case where the σ algebra G is generated by a single random ¡ variable Y , i.e., G σ(Y ). In this case, every G measurable random variable is a Borel function Æ ¡ of Y (exercise!), so E(X G ) is the unique Borel function h(Y ) (up to sets of probability zero) that j minimizes E(X h(Y ))2. The following exercise indicates that the special case where G σ(Y ) ¡ Æ for some real-valued random variable Y is in fact very general. Exercise 1. Show that if G is countably generated (that is, there is some countable collection of set B G such that G is the smallest σ algebra containing all of the sets B ) then there is a j 2 ¡ j G measurable real random variable Y such that G σ(Y ).
    [Show full text]
  • Conditional Expectation and Prediction Conditional Frequency Functions and Pdfs Have Properties of Ordinary Frequency and Density Functions
    Conditional expectation and prediction Conditional frequency functions and pdfs have properties of ordinary frequency and density functions. Hence, associated with a conditional distribution is a conditional mean. Y and X are discrete random variables, the conditional frequency function of Y given x is pY|X(y|x). Conditional expectation of Y given X=x is Continuous case: Conditional expectation of a function: Consider a Poisson process on [0, 1] with mean λ, and let N be the # of points in [0, 1]. For p < 1, let X be the number of points in [0, p]. Find the conditional distribution and conditional mean of X given N = n. Make a guess! Consider a Poisson process on [0, 1] with mean λ, and let N be the # of points in [0, 1]. For p < 1, let X be the number of points in [0, p]. Find the conditional distribution and conditional mean of X given N = n. We first find the joint distribution: P(X = x, N = n), which is the probability of x events in [0, p] and n−x events in [p, 1]. From the assumption of a Poisson process, the counts in the two intervals are independent Poisson random variables with parameters pλ and (1−p)λ (why?), so N has Poisson marginal distribution, so the conditional frequency function of X is Binomial distribution, Conditional expectation is np. Conditional expectation of Y given X=x is a function of X, and hence also a random variable, E(Y|X). In the last example, E(X|N=n)=np, and E(X|N)=Np is a function of N, a random variable that generally has an expectation Taken w.r.t.
    [Show full text]