The Central Limit Theorem (CLT) Says That If Ξ Is a Random Variable With

The Central Limit Theorem (CLT) Says That If Ξ Is a Random Variable With

The Central Limit Theorem (CLT) says that if » is a random variable with mean < » > and variance σ2, then for large N the sample mean » + » + ::: + » m = 1 2 N (1) N is approximately Gaussian with mean < » > and variance σ2=N. Let us take an example: let » be a Poisson random variable with mean 10. Then its variance is also 10 (a peculiar property of the Poisson distribution). If we let N be 20 in the sample mean above, then we expect the sample mean to be approximately Gaussian with mean 10 and variance σ2 = 10=20 = 0:5, so that 2σ2 = 1; thus the probability distribution of the sample means should be approximately the Gaussian distribution 1 2 P (x) = p e¡(x¡10) (2) ¼ In fact, though, m is a discrete random variable. What do we mean when we say that its distribution is described by the distribution function (or probability density) P(x)? This can only mean that when we look in some interval, the two agree pretty well. Now if we compute sample means many times and make a histogram of the results, we fill up “bins” that correspond to certain ranges of values. Since » takes integer values, and in finding the mean we divide by N=20, the actual values that can occur for m are things like ... 9.35, 9.40, 9.45, .... These could all go into a bin for values between 9 and 9.5. Suppose we sample m many times, say Nm times. Then what does the Gaussian distribution predict for the number of values in that bin? We would have to integrate over the interval, to get the probability for landing in the interval, and then multiply by Nm, to find the actual number that landed there. That is we predict Z 9:5 Nm P (x)dx ¼ NmP (9:25)(0:5) (3) 9 values in the interval. Notice that the bin size, 0.5, comes in here. Clearly if we chose the bins larger than 0.5, there would be more values in each bin (and fewer bins). This number should agree with the histogram. You should see the agreement if you sample m many times (Nm times), do a histogram, and plot on top of it NmP (x)∆x, where ∆x is the bin size in the histogram. We’ll see an example. Here is a possibly confusing point. The Poisson distribution itself looks Gaus- sian for large mean ¸. Notice that this is a completely different phenomenon! It means that if ¸ is large and you sample the distribution many times (Np times), and make a histogram, it will look like the Gaussian 1 ¡(x¡¸)2=2¸ Np p e (4) 2¼¸ (I’m assuming here that you make the bin size 1, since the Poisson distribution takes values 0, 1, 2, ..., i.e., each integer value gets its own bin.) Think why this is NOT the Central Limit Theorem. 1.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    1 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us