Chapter 9. Properties of Point Estimators and Methods of Estimation

Chapter 9. Properties of Point Estimators and Methods of Estimation

Chapter 9. Properties of Point Estimators and Methods of Estimation 9.1 Introduction 9.2 Relative Efficiency 9.3 Consistency 9.4 Sufficiency 9.5 The Rao-Blackwell Theorem and Minimum- Variance Unbiased Estimation 9.6 The Method of Moments 9.7 The Method of Maximum Likelihood 1 9.1 Introduction • Estimator θ^ = θ^n = θ^(Y1;:::;Yn) for θ : a function of n random samples, Y1;:::;Yn. • Sampling distribution of θ^ : a probability distribution of θ^ • Unbiased estimator, θ^ : E(θ^) − θ = 0. • Properties of θ^ : efficiency, consistency, sufficiency • Rao-Blackwell theorem : an unbiased esti- mator with small variance is a function of a sufficient statistic • Estimation method - Minimum-Variance Unbiased Estimation - Method of Moments - Method of Maximum Likelihood 2 9.2 Relative Efficiency • We would like to have an estimator with smaller bias and smaller variance : if one can find several unbiased estimators, we want to use an estimator with smaller vari- ance. • Relative efficiency (Def 9.1) Suppose θ^1 and θ^2 are two unbi- ased estimators for θ, with variances, V (θ^1) and V (θ^2), respectively. Then relative efficiency of θ^1 relative to θ^2, denoted eff(θ^1; θ^2), is defined to be the ra- V (θ^2) tio eff(θ^1; θ^2) = V (θ^1) - eff(θ^1; θ^2) > 1 : V (θ^2) > V (θ^1), and θ^1 is relatively more efficient than θ^2. - eff(θ^1; θ^2) < 1 : V (θ^2) < V (θ^1), and θ^2 is relatively more efficient than θ^1. 3 (Example) want to estimate the mean of a normal distribution. Let θ^med be the sample median and θ^mean be the sample mean. Then θ^mean is better than θ^med. Why? ^ 2 ^ ^ V (θmean) σ =n eff(θmed; θmean) = = 2 2 = :6366. V (θ^med) (1:2533) σ =n (Exercise) Let Y1;:::;Yn denote a random sample from a population with mean µ and σ2. Consider the following three estimates for µ : + + ^ = 1( + ), ^ = 1 + Y2 ::: Yn−1 + 1 , ^ = ¯. µ1 2 Y1 Y2 µ2 4Y1 2(n−2) 4Yn µ3 Y (a) Show that they are unbiased. (b) Find the efficiency ofµ ^3 relative toµ ^2 andµ ^1, respectively. 4 9.3 Consistency • Motivation - (Tossing a coin) A coin is tossed n times independently, and p is the (unknown) probability of resulting in heads. Suppose we are interested in the number of heads among n tosses, Y . (Q1) what is the distribution of Y ? Since p is unknown, considerp ^ =p ^n = Y=n. As n increases, the amount of information in the sample(here, the quality ofp ^n) also increases in the sense thatp ^n = Y=n should be very close to p as n ! 1. (Q2) How one can express \closeness" of p^n to p? Sincep ^n = Y=n is a statistic, consider the probability that j p^n − p j will be less than some arbitrary positive number as n in- creases : what is the value of P (j Y=n−p |≤ ) as n ! 1? 5 • Consistency and related theorems (Def 9.2) θ^n = θ^(Y1;:::;Yn) is said to be a p consistent estimator of θ(i.e., θ^n −! θ) if, for any > 0, lim P (j θ^n − θ |≤ ) = 1 or lim P (j θ^n − θ j> ) = 0 n!1 n!1 (note) \θ^n is a consistent estimator of θ" means \θ^n converges in probability to θ" (Thm 9.1) An unbiased θ^n for θ is a con- sistent estimator of θ if limn!1 V (θ^n) = 0. (Example 9.2) Let Y1;:::;Yn denote a ran- dom sample from a distribution with mean 2 µ and variance σ < 1. Show that Y¯n = 1 Pn n i=1 Yi is a consistent estimator of µ. 6 How about consistency of Y¯1 − Y¯2 for µ1 − µ2? ^ ^` (Thm 9.2) Suppose that θn and θn are con- sistent estimators of θ and θ`, respectively. Then, ^ ^` ` (a) θn+θn is a consistent estimator of θ+θ ^ ^` ` (b) θn ×θn is a consistent estimator of θ×θ ` ^ ^` (c) If θ 6= 0, θn=θn is a consistent estima- tor of θ/θ` (d) If g(·) is a real-valued function that is continuous at θ, then g(θ^n) is a consistent estimator of g(θ). (Example 9.3) Let Y1;:::;Yn denote a ran- dom sample from a distribution with finite 2 ` 4 ` E(Yi) = µ, E(Yi ) = µ2, and E(Yi ) = µ4. 2 1 Pn ¯ 2 Show that Sn = n−1 i=1(Yi − Yn ) is a 2 consistent estimator of σ = V (Yi). 7 (Question) Suppose Y1;:::;Yn is a random sample from any distribution with mean µ and known variance σ2. Then, in Section 8.6, we derived a large-sample confidence interval for µ with confidence coefficient approximately equal to 1 − α, p p ¯ ¯ [Y − zα/2(σ= n), Y + zα/2(σ= n)]. If σ2 is unknown and n is large, can one replace σ with Sn? Why? (Thm 9.3) Suppose that Un has a distribu- tion function that converges to a standard normal distribution function as n ! 1. If Wn converges in probability to 1, then the distribution function of Un=Wn converges to a standard normal distribution. (Example 9.4) 8 (Note) p · Example 9.4 implies that when n is large, n(Y¯n − µ)=Sn has approximately a standard normal distri- bution whatever is the form of the distribution from which the sample is taken. · We know from Chapter 7(p.359) that if the sample pis taken from a normal distribution and n is finite, n(Y¯n −µ)=Sn has a t distribution with n−1 degrees of freedom · This implies that if a large sample is takenp from a normal distribution, the distribution of n(Y¯n − µ)=Sn can be approximated by a standard normal distribution. In other words, if n gets large, then the number of degrees of freedom also gets large, and the t- distribution can be approximated by a standard nor- mal distribution (see Table 4 and 5 in pp.848-849). (Exercise) Suppose that Y ∼ b(n; p). Show that 1)p ^n = Y=n is an unbiased estimator of p p 2) (^pn −p)= p^nq^n=n converges to a standard normal distribution. 3) Derive a large-sampel confidence interval for p with confidence coefficient 1 − α 9 9.4 Sufficiency • We summarize the information in the sam- ple by using a statistic, g(Y1;:::;Yn) for the parameters of interest. (Example) Y¯ and S2 as the unbiased esti- mator for the population mean µ and vari- ance σ2 (Question) Does the process of summariz- ing the original set of n sample observa- tions to a few statistics, g1(Y1;:::;Yn);:::; gm(Y1;:::;Yn) retain all the information about the parameters of interest in n sample ob- servations? (Answer) There are methods to find statis- tics g1(Y1;:::;Yn); : : : ; gm(Y1;:::;Yn) sum- marizing all the information in a sample about the parameters of interest : we call such statistics sufficient statistics. 10 • Sufficient statistic (Def 9.3) Let Y1;:::;Yn denote a random sample from a probability distribution with unknown parameter θ. Then the statis- tic U = g(Y1;:::;Yn) is said to be suffi- cient for θ if the conditional distribution of Y1;:::;Yn, given U, does not depend on θ. (Example) Consider the outcomes of n independent trials of a binomial experiment, X1;:::;Xn where Xi ∼ b(1; p) and p = P (the i-th trial is a success). Pn Suppose we are given a value of Y = i=1 Xi(i.e., # of successes among n trials). If we know the value of Y , do we have to look at X1;:::;Xn or other functions of X1;:::;Xn in order to gain further information about p? Why? P (X1 = x1;:::;Xn = xn j Y = y) does not depend on p (i.e., as long as Y is known, there is no further information about p from other functions of X1;:::;Xn). We call Y a sufficient statistic for p. 11 • How to find a sufficient statistic · (Def 9.3) tells us how to check if a statis- tic is sufficient or not. · likelihood of the sample(Def 9.4) and fac- torization criterion(Thm 9.4) help find a sufficient statistic for θ (Def 9.4) Let y1; : : : ; yn be sample observa- tion taken on corresponding random vari- ables Y1;:::;Yn whose distribution depends on θ. Then the likelihood of obtaining the sample y1; : : : ; yn when the parameter is θ, L(θ) = L(y1; : : : ; yn j θ), is defined to be i) p(y1; : : : ; yn j θ) for discrete Y1;:::;Yn ii) f(y1; : : : ; yn j θ) for continuous Y1;:::;Yn · what L(θ1) > L(θ2) means for given Y1 = y1;:::;Yn = yn? 12 (Def 9.4 continued) If the set of random variables Y1;:::;Yn is a random sample from iid p(y j θ) or f(y j θ)(i.e., Y1;:::;Yn ∼ p(y j θ) or f(y j θ)), then the likelihood of the sample L(θ) = L(y1; : : : ; yn j θ) is i) for discrete Y1;:::;Yn; n Y p(y1; : : : ; yn j θ) = p(yi j θ) = p(y1 j θ) × · · · × p(yn j θ) i=1 ii) for continuous Y1;:::;Yn; n Y f(y1; : : : ; yn j θ) = f(yi j θ) = f(y1 j θ) × · · · × f(yn j θ) i=1 (Thm 9.4: factorization criterion) Let U be a statistic based on the random sample Y1;:::;Yn.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    24 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us