Lecture 2: Moments, Cumulants, and Scaling

Total Page:16

File Type:pdf, Size:1020Kb

Lecture 2: Moments, Cumulants, and Scaling Lecture 2: Moments, Cumulants, and Scaling Scribe: Ernst A. van Nierop (and Martin Z. Bazant) February 4, 2005 Handouts: • Historical excerpt from Hughes, Chapter 2: Random Walks and Random Flights. • Problem Set #1 1 The Position of a Random Walk 1.1 General Formulation Starting at the originX� 0= 0, if one takesN steps of size�xn, Δ then one ends up at the new position X� N . This new position is simply the sum of independent random�xn), variable steps (Δ N � X� N= Δ�xn. n=1 The set{ X� n} contains all the positions the random walker passes during its walk. If the elements of the{ setΔ�xn} are indeed independent random variables,{ X� n then} is referred to as a “Markov chain”. In a MarkovX� chain,N+1is independentX� ofnforn < N, but does depend on the previous positionX� N . LetPn(R�) denote the probability density function (PDF) for theR� of values the random variable X� n. Note that this PDF can be calculated for both continuous and discrete systems, where the discrete system is treated as a continuous system with Dirac delta functions at the allowed (discrete) values. Letpn(�r|R�) denote the PDF�x forn. Δ Note that in this more general notation we have included the possibility that�r (which is the value�xn) of depends Δ R� on. In Markov­chain jargon,pn(�r|R�) is referred to as the “transition probability”X� N fromto state stateX� N+1. Using the PDFs we just defined, let’s return to Bachelier’s equation from the first lecture. 1.2 Bachelier’s Equation ¿From the Markov assumption, the probability of finding a randomR� after walkerNsteps at depends on the probability of getting toR� −position �r afterN − 1 steps, and the probability of stepping�r, so that � d PN (R�) = pN (�rR| � − �r)PN−1(R� − �r) d �r. 1 M.Z. Bazant – 18.366 Random Walks and Diffusion – Lecture 2 2 Now, if we assume that the transition probability is translationallypn( invariant,�r|R�) =pn(�r) then (i.e.the step does not depend on the currentR�). positionThis reduces the integral to a convolution and gives us Bachelier’s equation � d PN (R�) = pN (�r)PN−1(R� − �r) d �r, or PN=pN∗ PN−1 where∗’ ‘ denotes the convolution. NotePN− that1is itself the result of a convolution, and we can writePNin terms of the PDFs of the steps··· N) taken and the (1 PDF of the initialP0 position,, so that PN=p1∗ p2∗ p3∗ · · · ∗ pN∗ P0. In what follows, we assumeP0= thatδ(�x), so the random walk always starts at the origin. 1.3 The Convolution Theorem Convolutions are best treated by Fourier transform, which turns them into simple products. For two functionsf(x) andg(x), the Convolution Theorem states that f�∗ g(�k) =fˆ(�k)ˆg(�k). To derive this result, we make use of the definition of the Fourier transform, ∞ ∞ � � �� � f�∗ g(�k) = e−ik·�x f(�x�)g(� x − � x�) dd� x� dd� x −∞ −∞ ∞ ∞ � �� � � = f(�x�) e−ik·�xg(�x − � x�) dd� x dd� x� −∞ −∞ where we have inverted the order of integration by assuming both integrals are absolutely convergent � (which follows from the common assumption,f, g ∈ L2). If we now�y= let� x − � x we obtain ∞ � ∞ � � � � � � f�∗ g(�k) = f(�x�) e−ik�ye−i k�x g(�y) dd� y dd� x −∞ −∞ =fˆ(�k)ˆg(�k). As described in the 2003 notes, the same method can be used to demonstrate the convolution theorem for discrete functions, � � � � � � fˆ(k)ˆg(k) = e−ikyf(y) e−ikzg(z) y z � � � � = e−ikx g(y)f(x − y) x y =f�∗ g(k), where all the terms with an exponentzhave equal been to collected on the right hand side of the second line. The key property used in both derivationse−ikye−ikz is=e that−ik(y+z). M.Z. Bazant – 18.366 Random Walks and Diffusion – Lecture 2 3 1.4 Exact Solution for the PDF of the Position Using the convolution theorem, we now write � � � � PˆN (k) =pˆ1(k)ˆp2(k) . pˆN (k) N � = pˆn(�k). n=1 � � N In the case of identical steps, thisP simplifiesˆN (k) = (ˆp1 to(k)) . Using the inverse transform, we find that the PDF of the position of the walker is given by ∞ N d � � � d �k P (�x) = eik·�x pˆ (�k) . N n (2π)d −∞ n=1 This is an exact solution, although the integral cannot be evaluated analytically in terms of ele­ mentary functions, except in a few special cases. (See Hughes, Ch. 2.) However, we are generally interested in the “long­time limit” of manyN→ steps, ∞, where the integral can be approxi­ mated by asymptotic analysis. This will be pursued in the next lecture. Here, we focus on scaling properties of the distribution with the number of steps. 2 Moments and Cumulants 2.1 Characteristic Functions � The Fourier transform of a PDF,PˆN ( suchk) for asX� N, is generally called a “characteristic function” in the probability literature. For random walks, especially on lattices, the characteristic function for the individual steps,p(�k), isˆ often referred to as the “structure function”. Note that � ∞ Pˆ(�0) = P (�x) dd� x = 1 −∞ due to the normalizationP (�x), of and that this is a maximum, since � � � |Pˆ(�k =� 0) | ≤ |e−ik·�xP (�x)| dd�x = P (� x) dd� x = 1. IfP (�x) is nonzero on a set of nonzero measure, then the complex exponential factor causes cancel­ lations in the integral, resulting|Pˆ(�k)| in< 1 for�k �=�0. 2.2 Moments In the 1­dimensional case= 1), (d the moments of a randomX variable,, are defined as m1 =�x� 2 m2 =�x � . n mn =�x � . M.Z. Bazant – 18.366 Random Walks and Diffusion – Lecture 2 4 To get a feeling for the significance of moments,m1is note just that the mean or expected value of the random variable.m When= 0, the second moment,m , allows us to define the “root­ 1 √ 2 mean­square width” of the distribution,m2. In a multi­dimensional setting1), (d the > moments become tensors, (j1j2...jn) mn =�xj1 xj2. xjn�, so thenthmoment is a tensor ofn with rankdncomponents. The off­diagonal components measure anisotropy, and generally the moment tensors describe the “shape” of the distribution. In probability, a characteristic functionPˆ(�k) is also often referred to as a “moment­generating function”, because it conveniently encodes the moments in its Taylor expansion around the origin. For example,d for= 1, we have ∞ n n � (−i) k mn Pˆ(k) = n! n=0 1 = 1 − im k − m k2. 1 2 2 since ∂nPˆ m =in (0). n ∂kn Ford > 1, the moment tensors are generated by ∂nPˆ (j1j2...jn) n � mn =i (0) ∂kj1 ∂kj2. ∂kjn and from the Taylor expansion, 1 Pˆ(�k) = 1 − im · �k − �k · m· �k +. 1 2 2 Note thatPˆ(�k) is analytic around the origin if and only if all moments existn< and∞ are finite (m for alln). This condition breaks down in cases of distributions with “fat tails”, to be discussed in subsequent lectures. 2.3 Cumulants Certain nonlinear combinations of moments, called “cumulants”, arise naturally when analyzing sums of independent random variables. Consider d � � d �k P (�x) = eik·�xPˆ(�k) (2π)d d � � � d �k = eik·�x+ψ(k) (2π)d where ψ(�k) = logPˆ(�k) is called the “cumulant generating function”, whose Taylor coefficients at the origin (analogous to the moments above) are the “cumulants”, which ared = scalars 1 and, for more generally, tensors M.Z. Bazant – 18.366 Random Walks and Diffusion – Lecture 2 5 ford > 1 ∂nψ n(j1,j2,...,jn) � i cn = (0) ∂kj1 ∂kj2. ∂kjn 1 ψ(�k) =−ic · �k − �k · c· �k +. 1 2 2 The cumulants can be expressed in terms of the moments by equating Taylor expansion coeffi­ cients in Eq. (2.3).d = For 1, we find c1 =m1 or ‘mean’ 2 2 c2 =m2− m1=σ or ‘variance’= (σ standard deviation) 3 c3 =m3− 3m1m2+ 2m1 or ‘skewness’ 2 2 4 c4 =m4− 3m2− 4m1m3+ 12m1m2− 6m1 or ‘kurtosis’. (Note that “skewness” and “kurtosis” are also sometimes used for the normalized cumulants, defined in the next lecture.) Although these expressions increase in complexity ratherlth quickly, the cumulant can be expressed in terms ofl moments the first in closed form via the determinant of an ‘almost’ lower triangular matrix, � m 1 0 0 0 0. 0 � � 1 � � m m 1 0 0 0. 0 � � 2 1 � � m m �2�m 1 0 0 . 0 � � 3 2 1 1 � � m m �3�m �3�m 1 0 . 0 � � 4 3 1 2 2 1 � l+1� �4� �4� �4� � cl=(−1) � m5 m4 m3 m2 m1 1 . 0 � . � 1 2 3 � � ...... ... .. .� � . .. � � . � � .. � � ml−1 ml−2 . 1 � � �l −1� � � ml ml−1 . .l− .2 . .m1� Ford >1, thenth cumulant is a tensorn ofwith rankdncomponents, related to the moment tensors,ml, for≤ 1l ≤ n. For example, the second cumulant matrix is given by (ij) (ij)(i)(j) c2 =m2 − m1m1. 3 Additivity of Cumulants A crucial feature of random walks with independently identically distributed (IID) steps is that � � cumulants are additive. If weψ( definek) andψN (k) to be the cumulant generating functions of the steps and the final position, respectively, then d � � ˆ � d �k P (�x) = eik·�x+N ψ(k) N (2π)d d � � � d �k = eik·�x+ψN(k) , (2π)d which implies � � ψN (k) =Nψ(k). M.Z. Bazant – 18.366 Random Walks and Diffusion – Lecture 2 6 th (j1,j2,...,jn) th Therefore, then cumulant tensorcn,N for the position afterNsteps is related ton the cumulant tensor for each step by cn,N=Ncn.
Recommended publications
  • 18.440: Lecture 9 Expectations of Discrete Random Variables
    18.440: Lecture 9 Expectations of discrete random variables Scott Sheffield MIT 18.440 Lecture 9 1 Outline Defining expectation Functions of random variables Motivation 18.440 Lecture 9 2 Outline Defining expectation Functions of random variables Motivation 18.440 Lecture 9 3 Expectation of a discrete random variable I Recall: a random variable X is a function from the state space to the real numbers. I Can interpret X as a quantity whose value depends on the outcome of an experiment. I Say X is a discrete random variable if (with probability one) it takes one of a countable set of values. I For each a in this countable set, write p(a) := PfX = ag. Call p the probability mass function. I The expectation of X , written E [X ], is defined by X E [X ] = xp(x): x:p(x)>0 I Represents weighted average of possible values X can take, each value being weighted by its probability. 18.440 Lecture 9 4 Simple examples I Suppose that a random variable X satisfies PfX = 1g = :5, PfX = 2g = :25 and PfX = 3g = :25. I What is E [X ]? I Answer: :5 × 1 + :25 × 2 + :25 × 3 = 1:75. I Suppose PfX = 1g = p and PfX = 0g = 1 − p. Then what is E [X ]? I Answer: p. I Roll a standard six-sided die. What is the expectation of number that comes up? 1 1 1 1 1 1 21 I Answer: 6 1 + 6 2 + 6 3 + 6 4 + 6 5 + 6 6 = 6 = 3:5. 18.440 Lecture 9 5 Expectation when state space is countable II If the state space S is countable, we can give SUM OVER STATE SPACE definition of expectation: X E [X ] = PfsgX (s): s2S II Compare this to the SUM OVER POSSIBLE X VALUES definition we gave earlier: X E [X ] = xp(x): x:p(x)>0 II Example: toss two coins.
    [Show full text]
  • Simulation of Moment, Cumulant, Kurtosis and the Characteristics Function of Dagum Distribution
    264 IJCSNS International Journal of Computer Science and Network Security, VOL.17 No.8, August 2017 Simulation of Moment, Cumulant, Kurtosis and the Characteristics Function of Dagum Distribution Dian Kurniasari1*,Yucky Anggun Anggrainy1, Warsono1 , Warsito2 and Mustofa Usman1 1Department of Mathematics, Faculty of Mathematics and Sciences, Universitas Lampung, Indonesia 2Department of Physics, Faculty of Mathematics and Sciences, Universitas Lampung, Indonesia Abstract distribution is a special case of Generalized Beta II The Dagum distribution is a special case of Generalized Beta II distribution. Dagum (1977, 1980) distribution has been distribution with the parameter q=1, has 3 parameters, namely a used in studies of income and wage distribution as well as and p as shape parameters and b as a scale parameter. In this wealth distribution. In this context its features have been study some properties of the distribution: moment, cumulant, extensively analyzed by many authors, for excellent Kurtosis and the characteristics function of the Dagum distribution will be discussed. The moment and cumulant will be survey on the genesis and on empirical applications see [4]. discussed through Moment Generating Function. The His proposals enable the development of statistical characteristics function will be discussed through the expectation of the definition of characteristics function. Simulation studies distribution used to empirical income and wealth data that will be presented. Some behaviors of the distribution due to the could accommodate both heavy tails in empirical income variation values of the parameters are discussed. and wealth distributions, and also permit interior mode [5]. Key words: A random variable X is said to have a Dagum distribution Dagum distribution, moment, cumulant, kurtosis, characteristics function.
    [Show full text]
  • Randomized Algorithms
    Chapter 9 Randomized Algorithms randomized The theme of this chapter is madendrizo algorithms. These are algorithms that make use of randomness in their computation. You might know of quicksort, which is efficient on average when it uses a random pivot, but can be bad for any pivot that is selected without randomness. Analyzing randomized algorithms can be difficult, so you might wonder why randomiza- tion in algorithms is so important and worth the extra effort. Well it turns out that for certain problems randomized algorithms are simpler or faster than algorithms that do not use random- ness. The problem of primality testing (PT), which is to determine if an integer is prime, is a good example. In the late 70s Miller and Rabin developed a famous and simple random- ized algorithm for the problem that only requires polynomial work. For over 20 years it was not known whether the problem could be solved in polynomial work without randomization. Eventually a polynomial time algorithm was developed, but it is much more complicated and computationally more costly than the randomized version. Hence in practice everyone still uses the randomized version. There are many other problems in which a randomized solution is simpler or cheaper than the best non-randomized solution. In this chapter, after covering the prerequisite background, we will consider some such problems. The first we will consider is the following simple prob- lem: Question: How many comparisons do we need to find the top two largest numbers in a sequence of n distinct numbers? Without the help of randomization, there is a trivial algorithm for finding the top two largest numbers in a sequence that requires about 2n − 3 comparisons.
    [Show full text]
  • Probability Cheatsheet V2.0 Thinking Conditionally Law of Total Probability (LOTP)
    Probability Cheatsheet v2.0 Thinking Conditionally Law of Total Probability (LOTP) Let B1;B2;B3; :::Bn be a partition of the sample space (i.e., they are Compiled by William Chen (http://wzchen.com) and Joe Blitzstein, Independence disjoint and their union is the entire sample space). with contributions from Sebastian Chiu, Yuan Jiang, Yuqi Hou, and Independent Events A and B are independent if knowing whether P (A) = P (AjB )P (B ) + P (AjB )P (B ) + ··· + P (AjB )P (B ) Jessy Hwang. Material based on Joe Blitzstein's (@stat110) lectures 1 1 2 2 n n A occurred gives no information about whether B occurred. More (http://stat110.net) and Blitzstein/Hwang's Introduction to P (A) = P (A \ B1) + P (A \ B2) + ··· + P (A \ Bn) formally, A and B (which have nonzero probability) are independent if Probability textbook (http://bit.ly/introprobability). Licensed and only if one of the following equivalent statements holds: For LOTP with extra conditioning, just add in another event C! under CC BY-NC-SA 4.0. Please share comments, suggestions, and errors at http://github.com/wzchen/probability_cheatsheet. P (A \ B) = P (A)P (B) P (AjC) = P (AjB1;C)P (B1jC) + ··· + P (AjBn;C)P (BnjC) P (AjB) = P (A) P (AjC) = P (A \ B1jC) + P (A \ B2jC) + ··· + P (A \ BnjC) P (BjA) = P (B) Last Updated September 4, 2015 Special case of LOTP with B and Bc as partition: Conditional Independence A and B are conditionally independent P (A) = P (AjB)P (B) + P (AjBc)P (Bc) given C if P (A \ BjC) = P (AjC)P (BjC).
    [Show full text]
  • Joint Probability Distributions
    ST 380 Probability and Statistics for the Physical Sciences Joint Probability Distributions In many experiments, two or more random variables have values that are determined by the outcome of the experiment. For example, the binomial experiment is a sequence of trials, each of which results in success or failure. If ( 1 if the i th trial is a success Xi = 0 otherwise; then X1; X2;:::; Xn are all random variables defined on the whole experiment. 1 / 15 Joint Probability Distributions Introduction ST 380 Probability and Statistics for the Physical Sciences To calculate probabilities involving two random variables X and Y such as P(X > 0 and Y ≤ 0); we need the joint distribution of X and Y . The way we represent the joint distribution depends on whether the random variables are discrete or continuous. 2 / 15 Joint Probability Distributions Introduction ST 380 Probability and Statistics for the Physical Sciences Two Discrete Random Variables If X and Y are discrete, with ranges RX and RY , respectively, the joint probability mass function is p(x; y) = P(X = x and Y = y); x 2 RX ; y 2 RY : Then a probability like P(X > 0 and Y ≤ 0) is just X X p(x; y): x2RX :x>0 y2RY :y≤0 3 / 15 Joint Probability Distributions Two Discrete Random Variables ST 380 Probability and Statistics for the Physical Sciences Marginal Distribution To find the probability of an event defined only by X , we need the marginal pmf of X : X pX (x) = P(X = x) = p(x; y); x 2 RX : y2RY Similarly the marginal pmf of Y is X pY (y) = P(Y = y) = p(x; y); y 2 RY : x2RX 4 / 15 Joint
    [Show full text]
  • Functions of Random Variables
    Names for Eg(X ) Generating Functions Topic 8 The Expected Value Functions of Random Variables 1 / 19 Names for Eg(X ) Generating Functions Outline Names for Eg(X ) Means Moments Factorial Moments Variance and Standard Deviation Generating Functions 2 / 19 Names for Eg(X ) Generating Functions Means If g(x) = x, then µX = EX is called variously the distributional mean, and the first moment. • Sn, the number of successes in n Bernoulli trials with success parameter p, has mean np. • The mean of a geometric random variable with parameter p is (1 − p)=p . • The mean of an exponential random variable with parameter β is1 /β. • A standard normal random variable has mean0. Exercise. Find the mean of a Pareto random variable. Z 1 Z 1 βαβ Z 1 αββ 1 αβ xf (x) dx = x dx = βαβ x−β dx = x1−β = ; β > 1 x β+1 −∞ α x α 1 − β α β − 1 3 / 19 Names for Eg(X ) Generating Functions Moments In analogy to a similar concept in physics, EX m is called the m-th moment. The second moment in physics is associated to the moment of inertia. • If X is a Bernoulli random variable, then X = X m. Thus, EX m = EX = p. • For a uniform random variable on [0; 1], the m-th moment is R 1 m 0 x dx = 1=(m + 1). • The third moment for Z, a standard normal random, is0. The fourth moment, 1 Z 1 z2 1 z2 1 4 4 3 EZ = p z exp − dz = −p z exp − 2π −∞ 2 2π 2 −∞ 1 Z 1 z2 +p 3z2 exp − dz = 3EZ 2 = 3 2π −∞ 2 3 z2 u(z) = z v(t) = − exp − 2 0 2 0 z2 u (z) = 3z v (t) = z exp − 2 4 / 19 Names for Eg(X ) Generating Functions Factorial Moments If g(x) = (x)k , where( x)k = x(x − 1) ··· (x − k + 1), then E(X )k is called the k-th factorial moment.
    [Show full text]
  • Expected Value and Variance
    Expected Value and Variance Have you ever wondered whether it would be \worth it" to buy a lottery ticket every week, or pondered questions such as \If I were offered a choice between a million dollars, or a 1 in 100 chance of a billion dollars, which would I choose?" One method of deciding on the answers to these questions is to calculate the expected earnings of the enterprise, and aim for the option with the higher expected value. This is a useful decision making tool for problems that involve repeating many trials of an experiment | such as investing in stocks, choosing where to locate a business, or where to fish. (For once-off decisions with high stakes, such as the choice between a sure 1 million dollars or a 1 in 100 chance of a billion dollars, it is unclear whether this is a useful tool.) Example John works as a tour guide in Dublin. If he has 200 people or more take his tours on a given week, he earns e1,000. If the number of tourists who take his tours is between 100 and 199, he earns e700. If the number is less than 100, he earns e500. Thus John has a variable weekly income. From experience, John knows that he earns e1,000 fifty percent of the time, e700 thirty percent of the time and e500 twenty percent of the time. John's weekly income is a random variable with a probability distribution Income Probability e1,000 0.5 e700 0.3 e500 0.2 Example What is John's average income, over the course of a 50-week year? Over 50 weeks, we expect that John will earn I e1000 on about 25 of the weeks (50%); I e700 on about 15 weeks (30%); and I e500 on about 10 weeks (20%).
    [Show full text]
  • Random Variables and Applications
    Random Variables and Applications OPRE 6301 Random Variables. As noted earlier, variability is omnipresent in the busi- ness world. To model variability probabilistically, we need the concept of a random variable. A random variable is a numerically valued variable which takes on different values with given probabilities. Examples: The return on an investment in a one-year period The price of an equity The number of customers entering a store The sales volume of a store on a particular day The turnover rate at your organization next year 1 Types of Random Variables. Discrete Random Variable: — one that takes on a countable number of possible values, e.g., total of roll of two dice: 2, 3, ..., 12 • number of desktops sold: 0, 1, ... • customer count: 0, 1, ... • Continuous Random Variable: — one that takes on an uncountable number of possible values, e.g., interest rate: 3.25%, 6.125%, ... • task completion time: a nonnegative value • price of a stock: a nonnegative value • Basic Concept: Integer or rational numbers are discrete, while real numbers are continuous. 2 Probability Distributions. “Randomness” of a random variable is described by a probability distribution. Informally, the probability distribution specifies the probability or likelihood for a random variable to assume a particular value. Formally, let X be a random variable and let x be a possible value of X. Then, we have two cases. Discrete: the probability mass function of X specifies P (x) P (X = x) for all possible values of x. ≡ Continuous: the probability density function of X is a function f(x) that is such that f(x) h P (x < · ≈ X x + h) for small positive h.
    [Show full text]
  • Expectation and Functions of Random Variables
    POL 571: Expectation and Functions of Random Variables Kosuke Imai Department of Politics, Princeton University March 10, 2006 1 Expectation and Independence To gain further insights about the behavior of random variables, we first consider their expectation, which is also called mean value or expected value. The definition of expectation follows our intuition. Definition 1 Let X be a random variable and g be any function. 1. If X is discrete, then the expectation of g(X) is defined as, then X E[g(X)] = g(x)f(x), x∈X where f is the probability mass function of X and X is the support of X. 2. If X is continuous, then the expectation of g(X) is defined as, Z ∞ E[g(X)] = g(x)f(x) dx, −∞ where f is the probability density function of X. If E(X) = −∞ or E(X) = ∞ (i.e., E(|X|) = ∞), then we say the expectation E(X) does not exist. One sometimes write EX to emphasize that the expectation is taken with respect to a particular random variable X. For a continuous random variable, the expectation is sometimes written as, Z x E[g(X)] = g(x) d F (x). −∞ where F (x) is the distribution function of X. The expectation operator has inherits its properties from those of summation and integral. In particular, the following theorem shows that expectation preserves the inequality and is a linear operator. Theorem 1 (Expectation) Let X and Y be random variables with finite expectations. 1. If g(x) ≥ h(x) for all x ∈ R, then E[g(X)] ≥ E[h(X)].
    [Show full text]
  • STAT 830 Generating Functions
    STAT 830 Generating Functions Richard Lockhart Simon Fraser University STAT 830 — Fall 2011 Richard Lockhart (Simon Fraser University) STAT 830 Generating Functions STAT 830 — Fall 2011 1 / 21 What I think you already have seen Definition of Moment Generating Function Basics of complex numbers Richard Lockhart (Simon Fraser University) STAT 830 Generating Functions STAT 830 — Fall 2011 2 / 21 What I want you to learn Definition of cumulants and cumulant generating function. Definition of Characteristic Function Elementary features of complex numbers How they “characterize” a distribution Relation to sums of independent rvs Richard Lockhart (Simon Fraser University) STAT 830 Generating Functions STAT 830 — Fall 2011 3 / 21 Moment Generating Functions pp 56-58 Def’n: The moment generating function of a real valued X is tX MX (t)= E(e ) defined for those real t for which the expected value is finite. Def’n: The moment generating function of X Rp is ∈ ut X MX (u)= E[e ] defined for those vectors u for which the expected value is finite. Formal connection to moments: ∞ k MX (t)= E[(tX ) ]/k! Xk=0 ∞ ′ k = µk t /k! . Xk=0 Sometimes can find power series expansion of MX and read off the moments of X from the coefficients of tk /k!. Richard Lockhart (Simon Fraser University) STAT 830 Generating Functions STAT 830 — Fall 2011 4 / 21 Moments and MGFs Theorem If M is finite for all t [ ǫ,ǫ] for some ǫ> 0 then ∈ − 1 Every moment of X is finite. ∞ 2 M isC (in fact M is analytic). ′ k 3 d µk = dtk MX (0).
    [Show full text]
  • 5.5 the Expected Value of a Function of Random Variables 5.6 Special
    Section 5. The Expected Value of a Function of Random Variables (ATTENDANCE 9)153 5.5 The Expected Value of a Function of Random Variables The material in this section is combined with the material in the next section and given in the next section. 5.6 Special Theorems Expected value of a function of random variables, E[g(Y ,Y ,...,Yk)], equals • 1 2 g(y1,y2,...,yk)p(y1,y2,...,yk) if discrete, = ∞all y1∞ all y2 ···∞ all yk P−∞ −∞P −∞ Pg(y1,y2,...,yk)f(y1,y2,...,yk) dy1dy2 dyk if continuous. ··· ··· R R R Also, the variance of a function of random variables, V [g(Y ,Y ,...,Yk)], with • 1 2 expected value, E[g(Y1,Y2,...,Yk)], is defined 2 V [g(Y ,Y ,...,Yk)] = E (g(Y ,Y ,...,Yk) E[g(Y ,Y ,...,Yk)]) 1 2 1 2 − 1 2 2 2 = E g(Y ,Y ,...,Yk) [E [g(Y ,Y ,...,Yk)]] . 1 2 − 1 2 Related to this, • E[Y + Y + + Yk] = E[Y ]+ + E[Yk], 1 2 ··· 1 ··· E[g (Y ,Y )+ + gk(Y ,Y )] = E[g (Y ,Y )] + + E[gk(Y ,Y )], 1 1 2 ··· 1 2 1 1 2 ··· 1 2 E(c) = c, if c is a constant, E[cg(Y1,Y2)] = cE[g(Y1,Y2)], if c is a constant, E[g(Y1)h(Y2)] = E[g(Y1)]E[h(Y2)], if g(Y1), h(Y2) independent. Exercise 5.6 (Special Theorems) 1. Discrete Expected Value Calculations: Waiting Times To Catch Fish. The joint density, p(y1,y2), of the number of minutes waiting to catch the first fish, y1, and the number of minutes waiting to catch the second fish, y2, is given below.
    [Show full text]
  • Moments & Cumulants
    ECON-UB 233 Dave Backus @ NYU Lab Report #1: Moments & Cumulants Revised: September 17, 2015 Due at the start of class. You may speak to others, but whatever you hand in should be your own work. Use Matlab where possible and attach your code to your answer. Solution: Brief answers follow, but see also the attached Matlab program. Down- load this document as a pdf, open it, and click on the pushpin: 1. Moments of normal random variables. This should be review, but will get you started with moments and generating functions. Suppose x is a normal random variable with mean µ = 0 and variance σ2. (a) What is x's standard deviation? (b) What is x's probability density function? (c) What is x's moment generating function (mgf)? (Don't derive it, just write it down.) (d) What is E(ex)? (e) Let y = a + bx. What is E(esy)? How does it tell you that y is normal? Solution: (a) The standard deviation is the (positive) square root of the variance, namely σ if σ > 0 (or jσj if you want to be extra precise). (b) The pdf is p(x) = (2πσ2)−1=2 exp(−x2=2σ2): (c) The mgf is h(s) = exp(s2σ2=2). (d) E(ex) is just the mgf evaluated at s = 1: h(1) = eσ2=2. (e) The mfg of y is sy s(a+bx) sa sbx sa hy(s) = E(e ) = E(e ) = e E(e ) = e hx(bs) 2 2 2 2 = esa+(bs) σ =2 = esa+s (bσ) =2: This has the form of a normal random variable with mean a (the coefficient of s) and variance (bσ)2 (the coefficient of s2=2).
    [Show full text]