
Lecture 10: Confidence intervals 1 of 16 Course: Mathematical Statistics Term: Fall 2017 Instructor: Gordan Žitkovi´c Lecture 10 Confidence intervals 10.1 Point vs. interval estimators We talked about point estimators in the previous lectures, and now we move on to interval estimators. Their main advantage over point estimators is that they give us an idea of how accurate our estimate is. Example 10.1.1. Consider the following two data sets, both consisting of measurements of the same quantity (say the distance to Proxima Centauri) and in the same units, but made with two different methods. method 1: 4.51, 4.52, 4.48, 4.49, 4.47, 4.53 method 2: 14.12, 1.30, 0.40, 2.50, 1.00, 3.18 If we use the sample mean Y¯ as the (point) estimator for the “true” mean m, both of these data sets yield the same result, namely Y¯ = 4.5. It is clear, however, that the first method is more accurate and that, in general, one should trust the results produced by method 1 more than those obtained by method 2. An interval estimator, i.e., a pair qˆL ≤ qˆR of point estimators, helps us convey this missing information. We usually think of qˆL and qˆR as end- points of an interval [qˆL, qˆR], called a confidence interval, with the property that the “true” parameter qˆ lies in [qˆLqˆR], “with high probability”. To be useful, confidence intervals should have the following properties 1. they should be short, and 2. they should contain q with high probability. These two, like in many other situations arising in statistics, cannot be opti- mized simultaneously. Indeed, as we make the interval shorter and shorter, the probability that it contains q will typically get smaller and smaller. On the other hand, we can make this probability get as close to 1 as we like by making the interval larger and larger. The usual way out of this (and we will see the same approach in the lecture on testing later) is to decide on the Last Updated: September 25, 2019 Lecture 10: Confidence intervals 2 of 16 least level of one of these criteria we can tolerate, and then optimize the other under this constraint. In this case, we first pick a number a 2 (0, 1), called the significance level, and require that the interval contains q with the probability at least 1 − a, i.e., P[qL ≤ q ≤ qR] ≥ 1 − a.(10.1.1) The number 1 − a is called the confidence1 level or (confidence coefficient). As before, when we defined unbiasedness, the inequality in (10.1.1) is as- sumed to hold for all values of the parameter q. Therefore, the goal is to construct the the shortest interval such that (10.1.1) still holds. Sometimes, you only care about not underestimating (or not overestimat- ing) the parameter, without worrying about overestimating (or underestimat- ing) it. In those cases we talk about one-sided confidence intervals and set either qL = −¥ or qR = ¥. In those cases, the requirement (10.1.1) is either P[q ≤ qR] ≥ 1 − a in the former, and P[qL ≤ q] ≥ 1 − a in the later. If the probabilities corresponding to both tails are the same, (and, typically, equal to 1 − a/2) we say that the interval is symmetric. 10.2 Pivotal quantities Unfortunately, we do not have enough mathematics at our disposal to give an algorithm that will produce a confidence interval in every conceivable model. What we can do is describe one particular construction which, when it works (and it works in many cases of interest) yields very good results. The concept we depend on is similar to the concept of an estimator, but with one notable difference: Definition 10.2.1. Given a random sample (Y1,..., Yn) from a distribu- tion D with an unknown parameter q, a pivotal quantity is a function of the data (Y1,..., Yn) and the parameter q, whose distribution does not depend on q. So, unlike in the case of an estimator, the dependence on parameters is allowed, but at the expense of a different requirement. This time, the distribution of the pivotal quantity cannot depend on the parameter at all. Example 10.2.2. The normal model: 1. N(m, 1): Let (Y1,..., Yn) be a random sample from N(m, 1), with an unknown mean m, but known variance 1. The sample mean Y¯ is an estimator, but it is not a pivotal quantity. Indeed, we have seen before, that its distribution is normal with mean m and variance 1/n. This clearly depends on m. 1confidence+significance=1 Last Updated: September 25, 2019 Lecture 10: Confidence intervals 3 of 16 On the other hand, Y¯ − m is not an estimator, but it is a pivotal quantity. Indeed, it is normally distributed with mean 0 and variance 1/n - a distribution which does not depend on m. If we multiply a pivotal quantity by a constant (which depends neither on the unknown parameterp m nor on the data) we still get a pivotal quantity. This way, n(Y¯ − m) is also a pivotal quantity. What is its distribution? 2. N(m, s): This time, the parameter q is two-dimensional, i.e., q = (m, s), i.e., both m and s are unknown. To construct a pivotal quan- tity, we start with Y¯ − m from above, but, since s is unknown, Y¯ − m is no longer a pivotal quantity in this new model. Indeed, its dis- N( p1 ) tribution is 0, n s , which clearly depends on s. To make it independent of q, we need to get rid of s, as well. This is easy in this case - we simply need to normalize by s: Y¯ − m s is a pivotal quantity because its distribution is N(0, p1 ). As above, p n it remains a pivotal quantity if we multiply it by n. As an added benefit, this way we get rid of the dependence on n in the distribu- tion since p Y¯ − m n ∼ N(0, 1). s 10.3 Confidence intervals from pivotal quantities We are now ready for the construction of confidence intervals with the confi- dence level 1 − a (significance level a). It takes three easy steps: 1. Pick a pivotal quantity U (it will always be given to you in this course). 2. Define a and b by a = qU(a/2) and b = qU(1 − a/2). Since the distribution of U does not depend on q, its quantiles do not either. Therefore, no matter what q happens to be, we have P[a ≤ U ≤ b] = 1 − P[U < a or U > b] = 1 − (a/2 + a/2) = 1 − a. Last Updated: September 25, 2019 Lecture 10: Confidence intervals 4 of 16 3. “Solve” the inequality a ≤ U ≤ b for q, i.e., rearrange it so that the only thing left in the middle is q. Give names qˆL and qˆR to whatever you get on the left and on the right, respectively. Two comments need to be made here. First, if we set a = −¥ and b = qU(1 − a) in 2. above, we would get a one-sided interval. Same for a = qU(a) and b = +¥. In fact, by picking a = qU(a1) and b = qU(1 − a2) with a1 + a2 = 1, we get a whole family of confidence intervals (but we will not consider them in this course). Second, all of this works for one-dimensional q only. If, like in the normal model with unknown m and s, the parameter q = (m, s) is two-dimensional (or higher), “solving” for q in 3. above makes no sense. What could be con- structed, but we do not pursue it here, is a so-called confidence region, i.e. a region in R2 (or Rn if you have n components of q) which contains q with probability at least 1 − a. Example 10.3.1. p 1. N(m, 1): We use the pivotal quantity U = n(Y¯ − m), as in Example 10.2.2 above. U has the standard normal distribution N(0, 1), and, therefore, we can use the quantiles a and b of the standard normal. For concreteness, let us pick 1 − a = 95% and using tables (if you must) or software (the R-command is qnorm) compute the values of a and b as a = −1.96 and b = 1.96. This way, p P[−1.96 ≤ n(Y¯ − m) ≤ 1.96] = 0.95 Since p P[−1.96 ≤ n(Y¯ − m) ≤ 1.96] = 0.95 if and only if p p P[−1.96/ n ≤ Y¯ − m ≤ 1.96/ n] = 0.95 if and only if p p P[−Y¯ − 1.96/ n ≤ −m ≤ −Y¯ + 1.96/ n] = 0.95, if and only if p p P[Y¯ − 1.96/ n ≤ m ≤ Y¯ + 1.96/ n] = 0.95. and we can set p p qˆL = Y¯ − 1.96/ n and qˆR = Y¯ + 1.96/ n, so that, as required, P[qˆL ≤ m ≤ qˆR] = 1 − a. If we wanted a one-sided interval, we would pick qL = −¥ and qR would be the same as above, except that now the 1 − a/2 = 0.97.5- quantile 1.96 of N(0, 1) should bep replaced by the 1 − a = 0.95- quantile 1.64, i.e., qR = Y¯ + 1.64/ n. Last Updated: September 25, 2019 Lecture 10: Confidence intervals 5 of 16 Finally, let us consider a toy data set from N(m, 1): 7.54, 10.32, 8.34, 10.46, 10.69, 10.61.
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages16 Page
-
File Size-