1 Hypothesis testing
A statistical test is a method of making a decision about one hypothesis (the null hypothesis) in comparison with another one (the alternative) using a sample of observations of known size. A statistical test is not a proof per se. Accepting the null hypothesis (H0) doesn’t mean it is true, but just that the available observations are not incompatible with this hypothesis, and that there is not enough evidence to favour the alternative hypothesis over the null hypothesis.
There are 3 steps:
1. First specify a Null Hypothesis, usually denoted H0, which describes a model of interest. Usually, we express H0 as a restricted version of a more general model. 2. Then, construct a test statistic, which is a random variable (because it is a function of other random variables) with two features:
(a) it has a known distribution under the Null Hypothesis (usually, nor- mal or chi-square, t or F). Its distribution is known either because we assume enough about the distribution of the model disturbances to get small-sample distributions, or we assume enough to get asymp- totic distributions. (b) this known distribution may depend on data, but not on parameters (this is called pivotality: a test statistic is pivotal if it satisfies this condition).
3. Check whether or not the sample value of the test statistic is very far out in its sampling distribution.
When we perform a test, we may end up rejecting the null hypothesis even though it is true. In this case we are committing the so-called Type I error. The probability of type I error is the significance level (or size) of the test. It is also possible that we fail to reject the null hypothesis even though it is false. In this case we are committing the so-called Type II error. The probability of type II error is the power of the test.
2 Testing under Normality Assumption 2.1 Properties of OLS estimators Suppose
Y = Xβ + ε, 2 ε ∼ N(0N , σ IN ),
1 and X is full rank with rank K. Then,
0 −1 0 2 0 −1 (i) βb = (X X) X Y ∼ N β, σ (X X) e0e (ii) ∼ χ2(N − K) σ2 e0e (iii) s2 = is an unbiased estimator of σ2 and is independent of βb, N − K where e = Y − βXb .
Proof. (i) The fact that the disturbances are independent mean-zero normals, 2 0 0 2 ε ∼ N(0, σ IN ), implies E [X ε] = 0K and E [εε ] = σ IN ,so the OLS estimator is still BLUE: h i E βb = β,
V (βb) = σ2(X0X)−1.
Write out βb as a function of ε as follows: βb = (X0X)−1X0Y = β + (X0X)−1X0ε is a linear combination of a normally distributed vector. Since, for any vector x ∼ N(µ, Σ), (a + Ax) ∼ N(a + µ, AΣA0) (See Kennedy’s All About Variances Appendix), we have