Descriptive Statistics Gives All of the Information We Need

Total Page:16

File Type:pdf, Size:1020Kb

Descriptive Statistics Gives All of the Information We Need Answers S b e the mean return for a random sample of 1. Let S b e the S&P return for a randomly chosen dayand 20 days. (a) E ( S )=E (S )=:00032 p p (b) SD(S )=SD(S )= n = :00859= 20 = :00192 (c) :005 :00032 S>:005) = P Z> = P (Z>2:438) = :00074 P ( :00192 Without doing any calculations we know that it is more likely that an individual day's return will b e greater than .005 than that the mean return will b e, since the standard deviation of S is smaller than that of S . 2.(a) The graphical output for descriptive statistics gives all of the information we need: Descriptive Statistics Variable: Freshman Anderson-Darling Normality Test A-Squared: 0.794 P-Value: 0.038 Mean 82.3778 StDev 9.3983 Variance 88.3276 Skewness -3.2E-01 Kurtosis -7.8E-01 62.5 70.0 77.5 85.0 92.5 N 90 Minimum 59.0000 1st Quartile 75.0000 Median 83.0000 3rd Quartile 90.0000 95% Confidence Interval for Mu Maximum 98.0000 95% Confidence Interval for Mu 80.4093 84.3462 80 81 82 83 84 85 95% Confidence Interval for Sigma 8.1973 11.0148 95% Confidence Interval for Median 95% Confidence Interval for Median 81.2373 85.0000 So, the 95% con dence interval for the true average freshman retention rate is (80.409, 84.346). The interval requires that the variable b e reasonably normally distributed, and that the sample b e a random sample from a larger p opulation. The latter requirement is met, and the former is not to o bad. (b) The prediction interval is then r p 1 89 X t 1+ =82:378 (1:987)(9:398) 1:0111 = 82:378 18:777 = (63:601; 101:155): s :025 90 Note that the upp er limit of the interval go es past the highest p ossible value of 100, which re ects the slight left tail in the distribution. The actual value for NYU, bytheway,was 86. c 1999, Je rey S. Simono (c) Here is a graphical representation of the exp enditure variable: Descriptive Statistics Variable: Expend. Anderson-Darling Normality Test A-Squared: 6.531 P-Value: 0.000 Mean 14676.7 StDev 9801.9 Variance 96077185 Skewness 3.09663 Kurtosis 13.6810 10000 25000 40000 55000 70000 N 91 Minimum 6093.0 1st Quartile 8742.0 Median 11999.0 3rd Quartile 16923.0 95% Confidence Interval for Mu Maximum 73967.0 95% Confidence Interval for Mu 12635.4 16718.1 10000 11000 12000 13000 14000 15000 16000 17000 95% Confidence Interval for Sigma 8555.4 11476.9 95% Confidence Interval for Median 95% Confidence Interval for Median 10486.7 13416.6 The con dence interval for average (over all researchuniversities) mean (over all students within the university) exp enditure per student is thus (12635, 16718). This interval is not very useful, however, since the exp enditure variable is very long right{tailed. (d) Given the long right tail in the distribution, a prediction interval must b e constructed in the correct scale. The waytodoitistologthevariable, construct the prediction interval, and then antilog. Here is a normal plot for the logged exp enditure variable, which lo oks more Gaussian (it's still not great, b eing right{tailed, but there's no stronger transformation than the log to try): c 1999, Je rey S. Simono Normal Probability Plot for Logged e 99 Mean: 4.10586 StDev: 0.214029 95 90 80 70 60 50 40 Percent 30 20 10 5 1 3.5 3.7 3.9 4.1 4.3 4.5 4.7 4.9 Data The prediction interval in the logged scale has the form r p 1 90 =4:106 (1:987)(:214) X t 1+ 1:011 = 4:106 :4276 = (3:678; 4:534): s :025 91 3:678 4:534 Antilogging this interval gives (10 ; 10 ), or (4764; 34198). The actual value for NYU was 24135. Note that this interval is much more reasonable than the one done in the original scale, whichis 14677 (1:987)(9802)(1:0055) = 14677 19584 = (4907; 34261): (e) In the sample 30 of the 90 institutions have freshman retention rate less than 80%, so a con dence interval for the probability that a researchuniversity has retention rate less than 80% is s r 1 2 1 p(1 p) 3 3 = (1:96) = :333 :0974: p z :025 n 3 90 The continuity{corrected interval adds 1=180 = :0056 to the p ortion of the interval, giving :333 :103. 3.(a) From the marginal distribution determined earlier, the con dence interval has the form s 203 203 1 203 362 362 (1:645) ; 362 362 c 1999, Je rey S. Simono or :561 :043 = (:518;:604). The continuity{corrected interval is :561 (:043 + :001) = (:517;:605). (b) Using the joint distribution of employment opp ortunities and salary size, and counting the num- b er of resp ondents who rated salary as more imp ortant than employment opp ortunities gives the con dence interval s 58 58 1 58 362 362 (1:645) ; 362 362 or :16 :032 = (:128;:192). The continuity{corrected interval is :16 (:032 + :001) = (:127;:193). 4.(a) Here is MINITAB output: Confidence Intervals Variable N Mean StDev SE Mean 95.0 % CI Weekly p 88 0.880 8.186 0.873 ( -0.855, 2.614) The con dence interval for true average price change includes zero, so the observed value is not signi cantly di erent from zero. The price change variable is reasonably Gaussian here. p (b) A prediction interval for the price change of a randomly chosen sto ckis:88 (1:988)(8:186) 1:011 = :88 16:363. The wide interval re ects the large variability of sto ck prices. (c) Here are values separated bymarket: Descriptive Statistics Variable Market N Mean Median Tr Mean StDev SE Mean Weekly p OTC 51 0.64 0.00 0.09 9.43 1.32 NYSE 24 -0.03 -0.85 -0.58 5.96 1.22 ASE 13 3.49 1.30 2.79 6.18 1.71 Variable Market Min Max Q1 Q3 Weekly p OTC -19.40 34.20 -5.00 4.40 NYSE -7.60 19.70 -4.55 1.38 ASE -4.10 18.80 -1.05 6.30 The con dence intervals for eachmarketareasfollows: p 51 = :64 2:65 OTC ::64 (2:009)(9:43)= p 24 = :03 2:52 NYSE : :03 (2:069)(5:96)= p ASE :3:49 (2:179)(6:18)= 13 = 3:49 3:73 All three intervals contain zero, so the average returns in eachwere not signi cantly di erentfrom zero. Note that the ASE interval was close, but didn't quite make it b ecause of the small sample size. A sample of size 15 with the same mean and standard deviation would havegiven an interval that did not contain zero. The distributions of price change within each group are reasonably Gaussian here. (d) Here are prediction intervals: OTC ::64 (2:009)(9:43)(1:01) = :64 19:134 NYSE : :03 (2:069)(5:96)(1:021) = :03 12:59 ASE :3:49 (2:179)(6:18)(1:038) = 3:49 13:978 Note that the intervals for NYSE and ASE are narrower than that for all of the sto cks together, which re ects the gains in treating the markets separately (despite the reduction in e ective sample size). c 1999, Je rey S. Simono 5.(a) The mean of X is just the p opulation mean , or .00039. p p (b) The standard deviation (or standard error) of X is = n,or:00721= 30 = :00132. (c) :01 :00039 P (X>:01) = P Z> :00132 = P (Z>7:28) 0 It is more likely that the sample mean return will b e less than .01 than an individual return, since the mean is much more tightly centered around the p opulation mean. In fact, the probabilityof individual return b eing less than .01 is ab out .91. 6. Before we do anything, we should recognize that we're assuming that our data are a random sample from a roughly Gaussian distribution. Normal plots for return and yield show that the while the yield variable is reasonably Gaussian (with one p ossibly unusual lowvalue), returns are not, b eing bimo dal. In fact, as the structure of the data in the le suggests, there are twowell{de ned subgroups here: no load funds, and load funds. Normal Probability Plot for Yield ML Estimates Mean: 4.75179 99 StDev: 0.501942 95 90 80 70 60 50 40 30 Percent 20 10 5 1 3 4 5 6 Data Normal Probability Plot for Return ML Estimates Mean: 4.22628 99 StDev: 2.22312 95 90 80 70 60 50 40 30 Percent 20 10 5 1 0 5 10 Data Stem-and-leaf of Return N = 78 Leaf Unit = 0.10 c 1999, Je rey S. Simono 1 -1 7 2 -0 4 3 0 5 7 1 5789 27 2 33333445556777888889 (20) 3 00112223344444467889 31 4 445689 25 5 36667 20 6 023456 14 7 023444489 5 8 11278 p (n1) (n1) X t s= n. Since n =78, t =1:99. (a) Wewant a con dence interval, which has the form =2 =2 Here is the answer from Minitab: Variable N Mean StDev SE Mean 95.0 % CI Return 78 4.226 2.238 0.253 ( 3.722, 4.731) The nonnormality of returns here is less worrying, since the sample size is somewhat large, so this interval mightbeokay.
Recommended publications
  • Linear Regression in SPSS
    Linear Regression in SPSS Data: mangunkill.sav Goals: • Examine relation between number of handguns registered (nhandgun) and number of man killed (mankill) • Model checking • Predict number of man killed using number of handguns registered I. View the Data with a Scatter Plot To create a scatter plot, click through Graphs\Scatter\Simple\Define. A Scatterplot dialog box will appear. Click Simple option and then click Define button. The Simple Scatterplot dialog box will appear. Click mankill to the Y-axis box and click nhandgun to the x-axis box. Then hit OK. To see fit line, double click on the scatter plot, click on Chart\Options, then check the Total box in the box labeled Fit Line in the upper right corner. 60 60 50 50 40 40 30 30 20 20 10 10 Killed of People Number Number of People Killed 400 500 600 700 800 400 500 600 700 800 Number of Handguns Registered Number of Handguns Registered 1 Click the target button on the left end of the tool bar, the mouse pointer will change shape. Move the target pointer to the data point and click the left mouse button, the case number of the data point will appear on the chart. This will help you to identify the data point in your data sheet. II. Regression Analysis To perform the regression, click on Analyze\Regression\Linear. Place nhandgun in the Dependent box and place mankill in the Independent box. To obtain the 95% confidence interval for the slope, click on the Statistics button at the bottom and then put a check in the box for Confidence Intervals.
    [Show full text]
  • CONFIDENCE Vs PREDICTION INTERVALS 12/2/04 Inference for Coefficients Mean Response at X Vs
    STAT 141 REGRESSION: CONFIDENCE vs PREDICTION INTERVALS 12/2/04 Inference for coefficients Mean response at x vs. New observation at x Linear Model (or Simple Linear Regression) for the population. (“Simple” means single explanatory variable, in fact we can easily add more variables ) – explanatory variable (independent var / predictor) – response (dependent var) Probability model for linear regression: 2 i ∼ N(0, σ ) independent deviations Yi = α + βxi + i, α + βxi mean response at x = xi 2 Goals: unbiased estimates of the three parameters (α, β, σ ) tests for null hypotheses: α = α0 or β = β0 C.I.’s for α, β or to predictE(Y |X = x0). (A model is our ‘stereotype’ – a simplification for summarizing the variation in data) For example if we simulate data from a temperature model of the form: 1 Y = 65 + x + , x = 1, 2,..., 30 i 3 i i i Model is exactly true, by construction An equivalent statement of the LM model: Assume xi fixed, Yi independent, and 2 Yi|xi ∼ N(µy|xi , σ ), µy|xi = α + βxi, population regression line Remark: Suppose that (Xi,Yi) are a random sample from a bivariate normal distribution with means 2 2 (µX , µY ), variances σX , σY and correlation ρ. Suppose that we condition on the observed values X = xi. Then the data (xi, yi) satisfy the LM model. Indeed, we saw last time that Y |x ∼ N(µ , σ2 ), with i y|xi y|xi 2 2 2 µy|xi = α + βxi, σY |X = (1 − ρ )σY Example: Galton’s fathers and sons: µy|x = 35 + 0.5x ; σ = 2.34 (in inches).
    [Show full text]
  • Choosing a Coverage Probability for Prediction Intervals
    Choosing a Coverage Probability for Prediction Intervals Joshua LANDON and Nozer D. SINGPURWALLA We start by noting that inherent to the above techniques is an underlying distribution (or error) theory, whose net effect Coverage probabilities for prediction intervals are germane to is to produce predictions with an uncertainty bound; the nor- filtering, forecasting, previsions, regression, and time series mal (Gaussian) distribution is typical. An exception is Gard- analysis. It is a common practice to choose the coverage proba- ner (1988), who used a Chebychev inequality in lieu of a spe- bilities for such intervals by convention or by astute judgment. cific distribution. The result was a prediction interval whose We argue here that coverage probabilities can be chosen by de- width depends on a coverage probability; see, for example, Box cision theoretic considerations. But to do so, we need to spec- and Jenkins (1976, p. 254), or Chatfield (1993). It has been a ify meaningful utility functions. Some stylized choices of such common practice to specify coverage probabilities by conven- functions are given, and a prototype approach is presented. tion, the 90%, the 95%, and the 99% being typical choices. In- deed Granger (1996) stated that academic writers concentrate KEY WORDS: Confidence intervals; Decision making; Filter- almost exclusively on 95% intervals, whereas practical fore- ing; Forecasting; Previsions; Time series; Utilities. casters seem to prefer 50% intervals. The larger the coverage probability, the wider the prediction interval, and vice versa. But wide prediction intervals tend to be of little value [see Granger (1996), who claimed 95% prediction intervals to be “embarass- 1.
    [Show full text]
  • STATS 305 Notes1
    STATS 305 Notes1 Art Owen2 Autumn 2013 1The class notes were beautifully scribed by Eric Min. He has kindly allowed his notes to be placed online for stat 305 students. Reading these at leasure, you will spot a few errors and omissions due to the hurried nature of scribing and probably my handwriting too. Reading them ahead of class will help you understand the material as the class proceeds. 2Department of Statistics, Stanford University. 0.0: Chapter 0: 2 Contents 1 Overview 9 1.1 The Math of Applied Statistics . .9 1.2 The Linear Model . .9 1.2.1 Other Extensions . 10 1.3 Linearity . 10 1.4 Beyond Simple Linearity . 11 1.4.1 Polynomial Regression . 12 1.4.2 Two Groups . 12 1.4.3 k Groups . 13 1.4.4 Different Slopes . 13 1.4.5 Two-Phase Regression . 14 1.4.6 Periodic Functions . 14 1.4.7 Haar Wavelets . 15 1.4.8 Multiphase Regression . 15 1.5 Concluding Remarks . 16 2 Setting Up the Linear Model 17 2.1 Linear Model Notation . 17 2.2 Two Potential Models . 18 2.2.1 Regression Model . 18 2.2.2 Correlation Model . 18 2.3 TheLinear Model . 18 2.4 Math Review . 19 2.4.1 Quadratic Forms . 20 3 The Normal Distribution 23 3.1 Friends of N (0; 1)...................................... 23 3.1.1 χ2 .......................................... 23 3.1.2 t-distribution . 23 3.1.3 F -distribution . 24 3.2 The Multivariate Normal . 24 3.2.1 Linear Transformations . 25 3.2.2 Normal Quadratic Forms .
    [Show full text]
  • Inference in Normal Regression Model
    Inference in Normal Regression Model Dr. Frank Wood Remember I We know that the point estimator of b1 is P(X − X¯ )(Y − Y¯ ) b = i i 1 P 2 (Xi − X¯ ) I Last class we derived the sampling distribution of b1, it being 2 N(β1; Var(b1))(when σ known) with σ2 Var(b ) = σ2fb g = 1 1 P 2 (Xi − X¯ ) I And we suggested that an estimate of Var(b1) could be arrived at by substituting the MSE for σ2 when σ2 is unknown. MSE SSE s2fb g = = n−2 1 P 2 P 2 (Xi − X¯ ) (Xi − X¯ ) Sampling Distribution of (b1 − β1)=sfb1g I Since b1 is normally distribute, (b1 − β1)/σfb1g is a standard normal variable N(0; 1) I We don't know Var(b1) so it must be estimated from data. 2 We have already denoted it's estimate s fb1g I Using this estimate we it can be shown that b − β 1 1 ∼ t(n − 2) sfb1g where q 2 sfb1g = s fb1g It is from this fact that our confidence intervals and tests will derive. Where does this come from? I We need to rely upon (but will not derive) the following theorem For the normal error regression model SSE P(Y − Y^ )2 = i i ∼ χ2(n − 2) σ2 σ2 and is independent of b0 and b1. I Here there are two linear constraints P ¯ ¯ ¯ (Xi − X )(Yi − Y ) X Xi − X b1 = = ki Yi ; ki = P(X − X¯ )2 P (X − X¯ )2 i i i i b0 = Y¯ − b1X¯ imposed by the regression parameter estimation that each reduce the number of degrees of freedom by one (total two).
    [Show full text]
  • Sieve Bootstrap-Based Prediction Intervals for Garch Processes
    SIEVE BOOTSTRAP-BASED PREDICTION INTERVALS FOR GARCH PROCESSES by Garrett Tresch A capstone project submitted in partial fulfillment of graduating from the Academic Honors Program at Ashland University April 2015 Faculty Mentor: Dr. Maduka Rupasinghe, Assistant Professor of Mathematics Additional Reader: Dr. Christopher Swanson, Professor of Mathematics ABSTRACT Time Series deals with observing a variable—interest rates, exchange rates, rainfall, etc.—at regular intervals of time. The main objectives of Time Series analysis are to understand the underlying processes and effects of external variables in order to predict future values. Time Series methodologies have wide applications in the fields of business where mathematics is necessary. The Generalized Autoregressive Conditional Heteroscedasic (GARCH) models are extensively used in finance and econometrics to model empirical time series in which the current variation, known as volatility, of an observation is depending upon the past observations and past variations. Various drawbacks of the existing methods for obtaining prediction intervals include: the assumption that the orders associated with the GARCH process are known; and the heavy computational time involved in fitting numerous GARCH processes. This paper proposes a novel and computationally efficient method for the creation of future prediction intervals using the Sieve Bootstrap, a promising resampling procedure for Autoregressive Moving Average (ARMA) processes. This bootstrapping technique remains efficient when computing future prediction intervals for the returns as well as the volatilities of GARCH processes and avoids extensive computation and parameter estimation. Both the included Monte Carlo simulation study and the exchange rate application demonstrate that the proposed method works very well under normal distributed errors.
    [Show full text]
  • On Small Area Prediction Interval Problems
    ASA Section on Survey Research Methods On Small Area Prediction Interval Problems Snigdhansu Chatterjee, Parthasarathi Lahiri, Huilin Li University of Minnesota, University of Maryland, University of Maryland Abstract In the small area context, prediction intervals are often pro- √ duced using the standard EBLUP ± zα/2 mspe rule, where Empirical best linear unbiased prediction (EBLUP) method mspe is an estimate of the true MSP E of the EBLUP and uses a linear mixed model in combining information from dif- zα/2 is the upper 100(1 − α/2) point of the standard normal ferent sources of information. This method is particularly use- distribution. These prediction intervals are asymptotically cor- ful in small area problems. The variability of an EBLUP is rect, in the sense that the coverage probability converges to measured by the mean squared prediction error (MSPE), and 1 − α for large sample size n. However, they are not efficient interval estimates are generally constructed using estimates of in the sense they have either under-coverage or over-coverage the MSPE. Such methods have shortcomings like undercover- problem for small n, depending on the particular choice of age, excessive length and lack of interpretability. We propose the MSPE estimator. In statistical terms, the coverage error a resampling driven approach, and obtain coverage accuracy of such interval is of the order O(n−1), which is not accu- of O(d3n−3/2), where d is the number of parameters and n rate enough for most applications of small area studies, many the number of observations. Simulation results demonstrate of which involve small n.
    [Show full text]
  • Bayesian Prediction Intervals for Assessing P-Value Variability in Prospective Replication Studies Olga Vsevolozhskaya1,Gabrielruiz2 and Dmitri Zaykin3
    Vsevolozhskaya et al. Translational Psychiatry (2017) 7:1271 DOI 10.1038/s41398-017-0024-3 Translational Psychiatry ARTICLE Open Access Bayesian prediction intervals for assessing P-value variability in prospective replication studies Olga Vsevolozhskaya1,GabrielRuiz2 and Dmitri Zaykin3 Abstract Increased availability of data and accessibility of computational tools in recent years have created an unprecedented upsurge of scientific studies driven by statistical analysis. Limitations inherent to statistics impose constraints on the reliability of conclusions drawn from data, so misuse of statistical methods is a growing concern. Hypothesis and significance testing, and the accompanying P-values are being scrutinized as representing the most widely applied and abused practices. One line of critique is that P-values are inherently unfit to fulfill their ostensible role as measures of credibility for scientific hypotheses. It has also been suggested that while P-values may have their role as summary measures of effect, researchers underappreciate the degree of randomness in the P-value. High variability of P-values would suggest that having obtained a small P-value in one study, one is, ne vertheless, still likely to obtain a much larger P-value in a similarly powered replication study. Thus, “replicability of P- value” is in itself questionable. To characterize P-value variability, one can use prediction intervals whose endpoints reflect the likely spread of P-values that could have been obtained by a replication study. Unfortunately, the intervals currently in use, the frequentist P-intervals, are based on unrealistic implicit assumptions. Namely, P-intervals are constructed with the assumptions that imply substantial chances of encountering large values of effect size in an 1234567890 1234567890 observational study, which leads to bias.
    [Show full text]
  • A Robust Rescaled Moment Test for Normality in Regression
    Journal of Mathematics and Statistics 5 (1): 54-62, 2009 ISSN 1549-3644 © 2009 Science Publications A Robust Rescaled Moment Test for Normality in Regression 1Md.Sohel Rana, 1Habshah Midi and 2A.H.M. Rahmatullah Imon 1Laboratory of Applied and Computational Statistics, Institute for Mathematical Research, University Putra Malaysia, 43400 Serdang, Selangor, Malaysia 2Department of Mathematical Sciences, Ball State University, Muncie, IN 47306, USA Abstract: Problem statement: Most of the statistical procedures heavily depend on normality assumption of observations. In regression, we assumed that the random disturbances were normally distributed. Since the disturbances were unobserved, normality tests were done on regression residuals. But it is now evident that normality tests on residuals suffer from superimposed normality and often possess very poor power. Approach: This study showed that normality tests suffer huge set back in the presence of outliers. We proposed a new robust omnibus test based on rescaled moments and coefficients of skewness and kurtosis of residuals that we call robust rescaled moment test. Results: Numerical examples and Monte Carlo simulations showed that this proposed test performs better than the existing tests for normality in the presence of outliers. Conclusion/Recommendation: We recommend using our proposed omnibus test instead of the existing tests for checking the normality of the regression residuals. Key words: Regression residuals, outlier, rescaled moments, skewness, kurtosis, jarque-bera test, robust rescaled moment test INTRODUCTION out that the powers of t and F tests are extremely sensitive to the hypothesized error distribution and may In regression analysis, it is a common practice over deteriorate very rapidly as the error distribution the years to use the Ordinary Least Squares (OLS) becomes long-tailed.
    [Show full text]
  • Testing Normality: a GMM Approach Christian Bontempsa and Nour
    Testing Normality: A GMM Approach Christian Bontempsa and Nour Meddahib∗ a LEEA-CENA, 7 avenue Edouard Belin, 31055 Toulouse Cedex, France. b D´epartement de sciences ´economiques, CIRANO, CIREQ, Universit´ede Montr´eal, C.P. 6128, succursale Centre-ville, Montr´eal (Qu´ebec), H3C 3J7, Canada. First version: March 2001 This version: June 2003 Abstract In this paper, we consider testing marginal normal distributional assumptions. More precisely, we propose tests based on moment conditions implied by normality. These moment conditions are known as the Stein (1972) equations. They coincide with the first class of moment conditions derived by Hansen and Scheinkman (1995) when the random variable of interest is a scalar diffusion. Among other examples, Stein equation implies that the mean of Hermite polynomials is zero. The GMM approach we adopt is well suited for two reasons. It allows us to study in detail the parameter uncertainty problem, i.e., when the tests depend on unknown parameters that have to be estimated. In particular, we characterize the moment conditions that are robust against parameter uncertainty and show that Hermite polynomials are special examples. This is the main contribution of the paper. The second reason for using GMM is that our tests are also valid for time series. In this case, we adopt a Heteroskedastic-Autocorrelation-Consistent approach to estimate the weighting matrix when the dependence of the data is unspecified. We also make a theoretical comparison of our tests with Jarque and Bera (1980) and OPG regression tests of Davidson and MacKinnon (1993). Finite sample properties of our tests are derived through a comprehensive Monte Carlo study.
    [Show full text]
  • 4.7 Confidence and Prediction Intervals
    4.7 Confidence and Prediction Intervals Instead of conducting tests we could find confidence intervals for a regression coefficient, or a set of regression coefficient, or for the mean of the response given values for the regressors, and sometimes we are interested in the range of possible values for the response given values for the regressors, leading to prediction intervals. 4.7.1 Confidence Intervals for regression coefficients We already proved that in the multiple linear regression model β^ is multivariate normal with mean β~ and covariance matrix σ2(X0X)−1, then is is also true that Lemma 1. p ^ For 0 ≤ i ≤ k βi is normally distributed with mean βi and variance σ Cii, if Cii is the i−th diagonal entry of (X0X)−1. Then confidence interval can be easily found as Lemma 2. For 0 ≤ i ≤ k a (1 − α) × 100% confidence interval for βi is given by È ^ n−p βi ± tα/2 σ^ Cii Continue Example. ?? 4 To find a 95% confidence interval for the regression coefficient first find t0:025 = 2:776, then p 2:38 ± 2:776(1:388) 0:018 $ 2:38 ± 0:517 The 95% confidence interval for β2 is [1.863, 2.897]. We are 95% confident that on average the blood pressure increases between 1.86 and 2.90 for every extra kilogram in weight after correcting for age. Joint confidence intervals for a set of regression coefficients. Instead of calculating individual confidence intervals for a number of regression coefficients one can find a joint confidence region for two or more regression coefficients at once.
    [Show full text]
  • Power Comparisons of Shapiro-Wilk, Kolmogorov-Smirnov, Lilliefors and Anderson-Darling Tests
    Journal ofStatistical Modeling and Analytics Vol.2 No.I, 21-33, 2011 Power comparisons of Shapiro-Wilk, Kolmogorov-Smirnov, Lilliefors and Anderson-Darling tests Nornadiah Mohd Razali1 Yap Bee Wah 1 1Faculty ofComputer and Mathematica/ Sciences, Universiti Teknologi MARA, 40450 Shah Alam, Selangor, Malaysia E-mail: nornadiah@tmsk. uitm. edu.my, yapbeewah@salam. uitm. edu.my ABSTRACT The importance of normal distribution is undeniable since it is an underlying assumption of many statistical procedures such as I-tests, linear regression analysis, discriminant analysis and Analysis of Variance (ANOVA). When the normality assumption is violated, interpretation and inferences may not be reliable or valid. The three common procedures in assessing whether a random sample of independent observations of size n come from a population with a normal distribution are: graphical methods (histograms, boxplots, Q-Q-plots), numerical methods (skewness and kurtosis indices) and formal normality tests. This paper* compares the power offour formal tests of normality: Shapiro-Wilk (SW) test, Kolmogorov-Smirnov (KS) test, Lillie/ors (LF) test and Anderson-Darling (AD) test. Power comparisons of these four tests were obtained via Monte Carlo simulation of sample data generated from alternative distributions that follow symmetric and asymmetric distributions. Ten thousand samples ofvarious sample size were generated from each of the given alternative symmetric and asymmetric distributions. The power of each test was then obtained by comparing the test of normality statistics with the respective critical values. Results show that Shapiro-Wilk test is the most powerful normality test, followed by Anderson-Darling test, Lillie/ors test and Kolmogorov-Smirnov test. However, the power ofall four tests is still low for small sample size.
    [Show full text]