Colin Cameron: Brief Asymptotic Theory for 240A
For 240A we do not go in to great detail. Key OLS results are in Section 1 and 4. The theorems cited in sections 2 and 3 are those from Appendix A of Cameron and Trivedi (2005), Microeconometrics: Methods and Applications.
1. A Roadmap for the OLS Estimator
1.1. Estimator and Model Assumptions 1 N 2 N Consider the OLS estimator = i=1 xi i=1 xiyi, where there is just one regressor and no intercept. We wantP to …nd theP properties of as sample size N . b ! 1 First, we need to make assumptions about the data generatingb process (dgp). We assume data are independent over i, the model is correctly speci…ed and the error is well behaved:
yi = xi + ui 2 ui xi [0; ], not necessarily normal. j 1 2 N 2 Then E[ ] = and V[ ] = i=1 xi for any N. b b P 1.2. Consistency of Intuitively collapses on its mean of as N , since V[ ] 0 as N . b ! 1 ! ! 1 The formal term is that converges in probability to , or that plim = , p or that b . We say that is consistent for . b ! The usual method ofb proof is not this simple as our toolkit works separatelyb on eachb average, and then combinesb results. Given the dgp, can be written as
1 N x u = + N i=1 i i : b 1 N x2 NP i=1 i The numerator can be shownb to go to zero,P by a law of large numbers, while the denominator goes to something nonzero. It follows that the ratio goes to zero. Formally:
1 N p plim x u 0 + N i=1 i i = + = : ! plim 1 N x2 plim 1 N x2 NP i=1 i N i=1 i b P P 1.3. Asymptotic Normality of Intuitively, if a central limit theorem can be applied, then b a a 2 N 2 1 [E[ ],V[ ]] [ , ( x ) ]; N N i=1 i a where means is “basymptoticallyb b distributedP as”. Again our toolkit works separately on each average, and then combines re- sults. The method is to rescale by pN, to get something with nondegenerate distribution, and
1 N 0; 2 plim 1 N x2 i=1 xiui N i=1 i pN( ) = pN d N 1 N 2 ! h 1 N 2 i N Pi=1 xi plim N i=1Pxi 1 b 1 d P 0; 2 plim N x2 P : !N N i=1 i " # P The key component is that 1 N x u has a normal distribution as N , pN i=1 i i by a central limit theorem. ! 1 Given this theoretical resultP we convert from pN( ) to and drop the plim, giving a 2 N 2 1 [ , ( x ) ]: b b N i=1 i P 2. Consistency b
The keys are (1) convergence in probability; (2) law of large numbers for an average; (3) combining pieces.
2.1. Convergence in Probability
Consider the limit behavior of a sequence of random variables bN as N . ! 1 This is a stochastic extension of a sequence of real numbers, such as aN = 2 + (3=N).
2 Examples include: (1) bN is an estimator, say ; (2) bN is a component of an 1 estimator, such as N i xiui; (3) bN is a test statistic. Due to sampling randomness we can never be certainb that a random sequence P bN , such as an estimator N , will be within a given small distance of its limit, even if the sample is in…nitely large. But we can be almost certain. Di¤erent ways of expressing this almost certaintyb correspond to di¤erent types of convergence of a sequence of random variables to a limit. The one most used in econometrics is convergence in probability. Recall that a sequence of nonstochastic real numbers aN converges to a if f g for any " > 0, there exists N = N (") such that for all N > N ,
aN a < ": j j e.g. if aN = 2+3=N; then the limit a = 2 since aN a = 2+3=N 2 = 3=N < " j j j j j j for all N > N = 3=": For a sequence of r.v.’swe cannot be certain that bN b < ", even for large N, due to the randomness. Instead, we require that thej probability j of being within " is arbitrarily close to one. Thus bN converges in probability to b if f g
lim Pr[ bN b < "] = 1; N !1 j j for any " > 0. A more formal de…nition is the following. De…nition A1: (Convergence in Probability) A sequence of random variables bN converges in probability to b if for any " > 0 and > 0, there exists f g N = N ("; ) such that for all N > N ,
Pr[ bN b < "] > 1 : j j
We write plim bN = b, where plim is short-hand for probability limit, or p bN b. The limit b may be a constant or a random variable. The usual de…nition of convergence! for a sequence of real variables is a special case of A1. For vector random variables we can apply the theory for each element of 2 bN . [Alternatively replace bN b by the scalar (bN b)0(bN b) = (b1N b1) + 2 j j + (bKN bK ) or by its square root bN b .] jj jj Now consider bN to be a sequence of parameter estimates . f g b 3 De…nition A2:(Consistency) An estimator is consistent for 0 if plim = : 0b Unbiasedness consistency. Unbiasedness states E[] = . Unbiasedness ; b 0 permits variability around 0 that need not disappear as the sample size goes to in…nity. b Consistency ; unbiasedness. e.g. add 1=N to an unbiased and consistent estimator - now biased but still consistent. A useful property of plim is that it can apply to transformations of random variables.
Theorem A3:(Probability Limit Continuity). Let bN be a …nite-dimensional vector of random variables, and g( ) be a real-valued function continuous at a constant vector point b. Then p p bN b g(bN ) g(b): ! ) ! This theorem is often referred to as Slutsky’s Theorem. We instead call Theo- rem A12 Slutsky’stheorem. Theorem A3 is one of the major reasons for the prevalence of asymptotic results versus …nite sample results in econometrics. It states a very convenient property that does not hold for expectations. For example, plim(aN ; bN ) = (a; b) implies plim(aN bN ) = ab, whereas E[aN bN ] generally di¤ers from E[a]E[b]. Similarly plim[aN =bN ] = a=b provided b = 0. There are several ways to establish convergence6 in probability. The brute force method uses De…nition A1, this is rarely done. It is often easier to establish alternative modes of convergence, notably convergence in mean square or use of Chebychev’sinequality, which in turn imply convergence in probability. But it is usually easiest to use a law of large numbers.
2.2. Laws of Large Numbers Laws of large numbers are theorems for convergence in probability (or almost surely) in the special case where the sequence bN is a sample average, i.e. f g bN = XN where 1 N X = X : N N i i=1 X 4 Note that Xi here is general notation for a random variable, and in the regression context does not necessarily denote the regressor variables. For example Xi = xiui. De…nition A7:(Law of Large Numbers)A weak law of large numbers (LLN) speci…es conditions on the individual terms Xi in XN under which p (XN E[XN ]) 0: ! For a strong law of large numbers the convergence is instead almost surely. If a LLN can be applied then
plim XN = lim E[XN ] in general 1 N = lim N i=1 E[Xi] if Xi independent over i = if Xi iid. P The simplest laws of large numbers assume that Xi is iid.
Theorem A8:(Kolmogorov LLN ) Let Xi be iid (independent and identically f g as distributed). If and only if E[Xi] = exists and E[ Xi ] < , then (XN ) 0. j j 1 ! The Kolmogorov LLN gives almost sure convergence. Usually convergence in probability is enough and we can use the weaker Khinchine’s Theorem.
Theorem A8b:(Khinchine’s Theorem) Let Xi be iid (independent and iden- f g p tically distributed). If and only if E[Xi] = exists, then (XN ) 0. ! There are other laws of large numbers. In particular, if Xi are independent but not identically distributed we can use the Markov LLN.
2.3. Consistency of OLS Estimator
1 N 1 N 2 Obtain the probability limit of = + [ N i=1 xiui]=[ N i=1 xi ], under simple random sampling with iid errors. Assume xi iid with mean x and ui iid with mean 0. b P P 1 p As xiui are iid, apply Khinchine’s Theorem yielding N i xiui E[xu] = E[x] E[u] = 0. ! 2 1 P 2 p 2 As xi are iid, apply Khinchine’sTheorem yielding N i xi E[x ] which we assume exists. ! P By Theorem A3 (Probability Limit Continuity) plim[aN =bN ] = a=b if b = 0. Then 6 plim 1 N x u 0 plim = + N i=1 i i = + = : plim 1 N x2 E[x2] NP i=1 i b P5 3. Asymptotic Normality
The keys are (1) convergence in distribution; (2) central limit theorem for an average; (3) combining pieces. Given consistency, the estimator has a degenerate distribution that collapses on 0 as N . So cannot do statistical inference. [Indeed there is no reason to ! 1 do it if N .] Need to magnify orb rescale to obtain a random variable with nondegenerate! 1 distribution as N . ! 1 b 3.1. Convergence in Distribution
Often the appropriate scale factor is pN, so consider bN = pN( 0). bN has an extremely complicated cumulative distribution function (cdf) FN . But like any other function FN it may have a limit function, where convergenceb is in the usual (nonstochastic) mathematical sense. De…nition A10:(Convergence in Distribution) A sequence of random variables bN is said to converge in distribution to a random variable b if f g
lim FN = F; N !1 at every continuity point of F , where FN is the distribution of bN , F is the distribution of b, and convergence is in the usual mathematical sense.
d We write bN b, and call F the limit distribution of bN . p ! d f g bN b implies bN b. ! ! d In general, the reverse is not true. But if b is a constant then bN b implies p ! bN b. ! To extend limit distribution to vector random variables simply de…ne FN and F to be the respective cdf’sof vectors bN and b. A useful property of convergence in distribution is that it can apply to trans- formations of random variables.
Theorem A11:(Limit Distribution Continuity). Let bN be a …nite-dimensional vector of random variables, and g( ) be a continuous real-valued function. Then d d bN b g(bN ) g(b): (3.1) ! ) !
6 This result is also called the Continuous Mapping Theorem.
d p Theorem A12:(Slutsky’sTheorem) If aN a and bN b, where a is a random variable and b is a constant, then ! !
d (i) aN + bN a + b d ! (ii) aN bN ab (3.2) !d (iii) aN =bN a=b, provided Pr[b = 0] = 0: ! Theorem A12 (also called Cramer’s Theorem) permits one to separately …nd the limit distribution of aN and the probability limit of bN , rather than having to consider the joint behavior of aN and bN . Result (ii) is especially useful and is sometimes called the Product Rule.
3.2. Central Limit Theorems Central limit theorems give convergence in distribution when the sequence bN is a sample average. A CLT is much easier way to get the plim than brute forcef g use of De…nition A10. By a LLN XN has a degenerate distribution as it converges to a constant, limE[XN ]. So scale (XN E[XN ]) by its standard deviation to construct a random variable with unit variance