Appendix Tables

Total Page:16

File Type:pdf, Size:1020Kb

Appendix Tables Appendix Tables 787 788 Appendix Tables Table A.1 Cumulative Binomial Probabilities Appendix Tables 789 Table A.1 Cumulative Binomial Probabilities (cont.) 790 Appendix Tables Table A.1 Cumulative Binomial Probabilities (cont.) Table A.2 Cumulative Poisson Probabilities Appendix Tables 791 Table A.2 Cumulative Poisson Probabilities (cont.) 792 Appendix Tables Table A.3 Standard Normal Curve Areas Appendix Tables 793 Table A.3 Standard Normal Curve Areas (cont.) 794 Appendix Tables Table A.4 The Incomplete Gamma Function Appendix Tables 795 Table A.5 Critical Values for t Distributions 796 Appendix Tables Table A.6 Critical Values for Chi-Squared Distributions Appendix Tables 797 Table A.7 t Curve Tail Areas 798 Appendix Tables Table A.7 t Curve Tail Areas (cont.) Appendix Tables 799 Table A.8 Critical Values for F Distributions 800 Appendix Tables Table A.8 Critical Values for F Distributions (cont.) Appendix Tables 801 Table A.8 Critical Values for F Distributions (cont.) 802 Appendix Tables Table A.8 Critical Values for F Distributions (cont.) Appendix Tables 803 Table A.8 Critical Values for F Distributions (cont.) 804 Appendix Tables Table A.8 Critical Values for F Distributions (cont.) Appendix Tables 805 Table A.9 Critical Values for Studentized Range Distributions 806 Appendix Tables Table A.10 Chi-Squared Curve Tail Areas Appendix Tables 807 Table A.10 Chi-Squared Curve Tail Areas (cont.) 808 Appendix Tables Table A.11 Critical Values for the Ryan–Joiner Test of Normality Appendix Tables 809 Table A.12 Critical Values for the Wilcoxon Signed-Rank Test 810 Appendix Tables Table A.13 Critical Values for the Wilcoxon Rank-Sum Test Appendix Tables 811 Table A.14 Critical Values for the Wilcoxon Signed-Rank Interval 812 Appendix Tables Table A.15 Critical Values for the Wilcoxon Rank-Sum Interval Appendix Tables Table A.16 b Curves for t Tests 813 Answers to Odd-Numbered Exercises Chapter 1 This display brings out the gap in the data: 1. a. Houston Chronicle, Des Moines Register, Chicago There are no scores in the high 70’s. Tribune, Washington Post b. Capital One, Campbell Soup, Merrill Lynch, 13. a. 2 23 Stem units: 1.0 Prudential 3 2344567789 Leaf units: .10 c. Bill Jasper, Kay Reinke, Helen Ford, David Menendez 4 01356889 d. 1.78, 2.44, 3.50, 3.04 5 00001114455666789 6 0000122223344456667789999 3. a. In a sample of 100 DVD players, what are the chances that more than 20 need service while under warranty? 7 00012233455555668 What are the chances that none need service while still 8 02233448 under warranty? 9 012233335666788 b. What proportion of all DVD players of this brand and 10 2344455688 model will need service within the warranty period? 11 2335999 12 37 5. a. No, the relevant conceptual population is all scores of all students who participate in the SI in conjunction 13 8 with this particular statistics course. 14 36 b. The advantage of randomly allocating students to the 15 0035 two groups is that the two groups should then be fairly 16 comparable before the study. If the two groups perform 17 differently in the class, we might attribute this to the 18 9 treatments (SI and control). If it were left to students to choose, stronger or more dedicated students might b. A representative value could be the median, 7.0. gravitate toward SI, confounding the results. c. The data appear to be highly concentrated, except for c. If all students were put in the treatment group there a few values on the positive side. would be no results with which to compare the d. No, there is skewness to the right, or positive treatments. skewness. e. The value 18.9 appears to be an outlier, being more 7. One could generate a simple random sample of all single than two stem units from the previous value. family homes in the city, or a stratified random sample by taking a simple random sample from each of the ten 15. a. district neighborhoods. From each of the homes in the Relative sample the necessary data would be collected. This Number frequency would be an enumerative study because there exists a finite, identifiable population of objects from which to nonconforming Frequency (Freq/60) sample. 0 7 0.117 9. a. There could be several explanations for the variability 1 12 0.200 of the measurements. Among them could be measuring 2 13 0.217 error, (due to mechanical or technical changes across 3 14 0.233 measurements), recording error, differences in 4 6 0.100 weather conditions at time of measurements, etc. 5 3 0.050 This study involves a conceptual population. There is b. 6 3 0.050 no sampling frame. 7 1 0.017 11. 6 l 034 810.017 6 h 667899 1.001 7 l 00122244 Doesn’t add exactly to 1 because relative 7 h Stem ¼ tens frequencies have been rounded 8 l 001111122344 Leaf ¼ ones 8 h 5557899 9l 03 9h 58 814 Chapter 1 815 b. .917, .867, 1 À.867 ¼ .133 c. The center of the histogram is somewhere around b. Class Freq Rel freq Density 2 or 3 and it shows that there is some positive skewness in the data. 0– < 50 8 0.08 .0016 50– < 100 13 0.13 .0026 17. a. .375 100– < 150 11 0.11 .0022 .218 b. < c. .242 150– 200 21 0.21 .0042 < d. The histogram is very positively skewed. 200– 300 26 0.26 .0026 300– < 400 12 0.12 .0012 19. a. The number of subdivisions having no cul-de-sacs is 400– < 500 4 0.04 .0004 ¼ 17/47 .362, or 36.2%. The proportion having at 500– < 600 3 0.03 .0003 least one cul-de-sac is 30/47 ¼ .638, or 63.8%. 600– < 900 2 0.02 .00007 100 1.00 y: Count Percent 0 17 36.17 c. .79 1 22 46.81 2 6 12.77 23. Class Freq Class Freq 3 1 2.13 5 1 2.13 10– < 20 8 1.1– < 1.2 2 N ¼ 47 20– < 30 14 1.2– < 1.3 6 .362, .638 30– < 40 8 1.3– < 1.4 7 40– < 50 4 1.4– < 1.5 9 50– < 60 3 1.5– < 1.6 6 b. z: Count Percent 60– < 70 2 1.6– < 1.7 4 70– < 80 1 1.7– < 1.8 5 0 13 27.66 40 1.8– < 1.9 1 1 11 23.40 40 2 3 6.38 3 7 14.89 4 5 10.64 The original distribution is positively skewed. 5 3 6.38 The transformation creates a much more symmetric, 6 3 6.38 mound-shaped histogram. 8 2 4.26 N ¼ 47 25. a. Class interval Freq Rel. Freq. .894, .830 0–< 50 9 0.18 21. a. 50–< 100 19 0.38 Class Freq Rel freq 100–< 150 11 0.22 150–< 200 4 0.08 < 0– 100 21 0.21 200–< 250 2 0.04 < 100– 200 32 0.32 250–< 300 2 0.04 < 200– 300 26 0.26 300–< 350 1 0.02 < 300– 400 12 0.12 350–< 400 1 0.02 < 400– 500 4 0.04 > ¼ 400 1 0.02 < 500– 600 3 0.03 50 1.00 600– < 700 1 0.01 700– < 800 0 0.00 800– < 900 1 0.01 100 1.00 The distribution is skewed to the right, or positively skewed. There is a gap in the histogram, and what appears to be an outlier in the ‘500–550’ interval. The histogram is skewed right, with a majority of observations between 0 and 300 cycles. The class holding the most observations is between 100 and 200 cycles. 816 Chapter 1 decreased to any value at least 370 without changing b. Class interval Freq. Rel. Freq. the median. d. 6.18 min; 6.16 min 2.25 < 2.75 2 0.04 2.75 < 3.25 2 0.04 35. a. 125. 3.25 < 3.75 3 0.06 b. If 127.6 is reported as 130, then the median is 130, a substantial change. When there is rounding or grouping, 3.75 < 4.25 8 0.16 the median can be highly sensitive to a small change. 4.25 < 4.75 18 0.36 4.75 < 5.25 10 0.20 37. x~ ¼ 92; xtrð25Þ ¼ 95:07; xtrð10Þ ¼ 102:23; x ¼ 119:3 5.25 < 5.75 4 0.08 Positive skewness causes the mean to be larger than the 5.75 < 6.25 3 0.06 median. Trimming moves the mean closer to the median. 39. a. y ¼ x þ c, y~ ¼ x~ þ c The distribution of the natural logs of the original data b. y ¼ cx, y~ ¼ cx~ is much more symmetric than the original. 41. a. 25.8 b. 49.31 c. 7.02 d. 49.31 c. .56, .14. 43. a. 2887.6, 2888 b. 7060.3 29. d. The frequency distribution is: 45. 24.36 Relative Relative 47. $1,961,160 Class frequency Class frequency 49. À3.5; 1.3, 1.9, 2.0, 2.3, À2.5 0 < 150 .193 900 < 1050 .019 51. a. 1, 6, 5 150 < 300 .183 1050 < 1200 .029 b. The box plot shows positive skewness. The two 300 < 450 .251 1200 < 1350 .005 longest runs are extreme outliers. 450 < 600 .148 1350 < 1500 .004 c. outlier: greater than 13.5 or less than À6.5 À 600 < 750 .097 1500 < 1650 .001 extreme outlier: greater than 21 or less than 14 d.
Recommended publications
  • Concentration and Consistency Results for Canonical and Curved Exponential-Family Models of Random Graphs
    CONCENTRATION AND CONSISTENCY RESULTS FOR CANONICAL AND CURVED EXPONENTIAL-FAMILY MODELS OF RANDOM GRAPHS BY MICHAEL SCHWEINBERGER AND JONATHAN STEWART Rice University Statistical inference for exponential-family models of random graphs with dependent edges is challenging. We stress the importance of additional structure and show that additional structure facilitates statistical inference. A simple example of a random graph with additional structure is a random graph with neighborhoods and local dependence within neighborhoods. We develop the first concentration and consistency results for maximum likeli- hood and M-estimators of a wide range of canonical and curved exponential- family models of random graphs with local dependence. All results are non- asymptotic and applicable to random graphs with finite populations of nodes, although asymptotic consistency results can be obtained as well. In addition, we show that additional structure can facilitate subgraph-to-graph estimation, and present concentration results for subgraph-to-graph estimators. As an ap- plication, we consider popular curved exponential-family models of random graphs, with local dependence induced by transitivity and parameter vectors whose dimensions depend on the number of nodes. 1. Introduction. Models of network data have witnessed a surge of interest in statistics and related areas [e.g., 31]. Such data arise in the study of, e.g., social networks, epidemics, insurgencies, and terrorist networks. Since the work of Holland and Leinhardt in the 1970s [e.g., 21], it is known that network data exhibit a wide range of dependencies induced by transitivity and other interesting network phenomena [e.g., 39]. Transitivity is a form of triadic closure in the sense that, when a node k is connected to two distinct nodes i and j, then i and j are likely to be connected as well, which suggests that edges are dependent [e.g., 39].
    [Show full text]
  • Use of Statistical Tables
    TUTORIAL | SCOPE USE OF STATISTICAL TABLES Lucy Radford, Jenny V Freeman and Stephen J Walters introduce three important statistical distributions: the standard Normal, t and Chi-squared distributions PREVIOUS TUTORIALS HAVE LOOKED at hypothesis testing1 and basic statistical tests.2–4 As part of the process of statistical hypothesis testing, a test statistic is calculated and compared to a hypothesised critical value and this is used to obtain a P- value. This P-value is then used to decide whether the study results are statistically significant or not. It will explain how statistical tables are used to link test statistics to P-values. This tutorial introduces tables for three important statistical distributions (the TABLE 1. Extract from two-tailed standard Normal, t and Chi-squared standard Normal table. Values distributions) and explains how to use tabulated are P-values corresponding them with the help of some simple to particular cut-offs and are for z examples. values calculated to two decimal places. STANDARD NORMAL DISTRIBUTION TABLE 1 The Normal distribution is widely used in statistics and has been discussed in z 0.00 0.01 0.02 0.03 0.050.04 0.05 0.06 0.07 0.08 0.09 detail previously.5 As the mean of a Normally distributed variable can take 0.00 1.0000 0.9920 0.9840 0.9761 0.9681 0.9601 0.9522 0.9442 0.9362 0.9283 any value (−∞ to ∞) and the standard 0.10 0.9203 0.9124 0.9045 0.8966 0.8887 0.8808 0.8729 0.8650 0.8572 0.8493 deviation any positive value (0 to ∞), 0.20 0.8415 0.8337 0.8259 0.8181 0.8103 0.8206 0.7949 0.7872 0.7795 0.7718 there are an infinite number of possible 0.30 0.7642 0.7566 0.7490 0.7414 0.7339 0.7263 0.7188 0.7114 0.7039 0.6965 Normal distributions.
    [Show full text]
  • Effect of Probability Distribution of the Response Variable in Optimal Experimental Design with Applications in Medicine †
    mathematics Article Effect of Probability Distribution of the Response Variable in Optimal Experimental Design with Applications in Medicine † Sergio Pozuelo-Campos *,‡ , Víctor Casero-Alonso ‡ and Mariano Amo-Salas ‡ Department of Mathematics, University of Castilla-La Mancha, 13071 Ciudad Real, Spain; [email protected] (V.C.-A.); [email protected] (M.A.-S.) * Correspondence: [email protected] † This paper is an extended version of a published conference paper as a part of the proceedings of the 35th International Workshop on Statistical Modeling (IWSM), Bilbao, Spain, 19–24 July 2020. ‡ These authors contributed equally to this work. Abstract: In optimal experimental design theory it is usually assumed that the response variable follows a normal distribution with constant variance. However, some works assume other probability distributions based on additional information or practitioner’s prior experience. The main goal of this paper is to study the effect, in terms of efficiency, when misspecification in the probability distribution of the response variable occurs. The elemental information matrix, which includes information on the probability distribution of the response variable, provides a generalized Fisher information matrix. This study is performed from a practical perspective, comparing a normal distribution with the Poisson or gamma distribution. First, analytical results are obtained, including results for the linear quadratic model, and these are applied to some real illustrative examples. The nonlinear 4-parameter Hill model is next considered to study the influence of misspecification in a Citation: Pozuelo-Campos, S.; dose-response model. This analysis shows the behavior of the efficiency of the designs obtained in Casero-Alonso, V.; Amo-Salas, M.
    [Show full text]
  • The Beta-Binomial Distribution Introduction Bayesian Derivation
    In Danish: 2005-09-19 / SLB Translated: 2008-05-28 / SLB Bayesian Statistics, Simulation and Software The beta-binomial distribution I have translated this document, written for another course in Danish, almost as is. I have kept the references to Lee, the textbook used for that course. Introduction In Lee: Bayesian Statistics, the beta-binomial distribution is very shortly mentioned as the predictive distribution for the binomial distribution, given the conjugate prior distribution, the beta distribution. (In Lee, see pp. 78, 214, 156.) Here we shall treat it slightly more in depth, partly because it emerges in the WinBUGS example in Lee x 9.7, and partly because it possibly can be useful for your project work. Bayesian Derivation We make n independent Bernoulli trials (0-1 trials) with probability parameter π. It is well known that the number of successes x has the binomial distribution. Considering the probability parameter π unknown (but of course the sample-size parameter n is known), we have xjπ ∼ bin(n; π); where in generic1 notation n p(xjπ) = πx(1 − π)1−x; x = 0; 1; : : : ; n: x We assume as prior distribution for π a beta distribution, i.e. π ∼ beta(α; β); 1By this we mean that p is not a fixed function, but denotes the density function (in the discrete case also called the probability function) for the random variable the value of which is argument for p. 1 with density function 1 p(π) = πα−1(1 − π)β−1; 0 < π < 1: B(α; β) I remind you that the beta function can be expressed by the gamma function: Γ(α)Γ(β) B(α; β) = : (1) Γ(α + β) In Lee, x 3.1 is shown that the posterior distribution is a beta distribution as well, πjx ∼ beta(α + x; β + n − x): (Because of this result we say that the beta distribution is conjugate distribution to the binomial distribution.) We shall now derive the predictive distribution, that is finding p(x).
    [Show full text]
  • 9. Binomial Distribution
    J. K. SHAH CLASSES Binomial Distribution 9. Binomial Distribution THEORETICAL DISTRIBUTION (Exists only in theory) 1. Theoretical Distribution is a distribution where the variables are distributed according to some definite mathematical laws. 2. In other words, Theoretical Distributions are mathematical models; where the frequencies / probabilities are calculated by mathematical computation. 3. Theoretical Distribution are also called as Expected Variance Distribution or Frequency Distribution 4. THEORETICAL DISTRIBUTION DISCRETE CONTINUOUS BINOMIAL POISSON Distribution Distribution NORMAL OR Student’s ‘t’ Chi-square Snedecor’s GAUSSIAN Distribution Distribution F-Distribution Distribution A. Binomial Distribution (Bernoulli Distribution) 1. This distribution is a discrete probability Distribution where the variable ‘x’ can assume only discrete values i.e. x = 0, 1, 2, 3,....... n 2. This distribution is derived from a special type of random experiment known as Bernoulli Experiment or Bernoulli Trials , which has the following characteristics (i) Each trial must be associated with two mutually exclusive & exhaustive outcomes – SUCCESS and FAILURE . Usually the probability of success is denoted by ‘p’ and that of the failure by ‘q’ where q = 1-p and therefore p + q = 1. (ii) The trials must be independent under identical conditions. (iii) The number of trial must be finite (countably finite). (iv) Probability of success and failure remains unchanged throughout the process. : 442 : J. K. SHAH CLASSES Binomial Distribution Note 1 : A ‘trial’ is an attempt to produce outcomes which is neither sure nor impossible in nature. Note 2 : The conditions mentioned may also be treated as the conditions for Binomial Distributions. 3. Characteristics or Properties of Binomial Distribution (i) It is a bi parametric distribution i.e.
    [Show full text]
  • Chapter 6 Continuous Random Variables and Probability
    EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2019 Chapter 6 Continuous Random Variables and Probability Distributions Chap 6-1 Probability Distributions Probability Distributions Ch. 5 Discrete Continuous Ch. 6 Probability Probability Distributions Distributions Binomial Uniform Hypergeometric Normal Poisson Exponential Chap 6-2/62 Continuous Probability Distributions § A continuous random variable is a variable that can assume any value in an interval § thickness of an item § time required to complete a task § temperature of a solution § height in inches § These can potentially take on any value, depending only on the ability to measure accurately. Chap 6-3/62 Cumulative Distribution Function § The cumulative distribution function, F(x), for a continuous random variable X expresses the probability that X does not exceed the value of x F(x) = P(X £ x) § Let a and b be two possible values of X, with a < b. The probability that X lies between a and b is P(a < X < b) = F(b) -F(a) Chap 6-4/62 Probability Density Function The probability density function, f(x), of random variable X has the following properties: 1. f(x) > 0 for all values of x 2. The area under the probability density function f(x) over all values of the random variable X is equal to 1.0 3. The probability that X lies between two values is the area under the density function graph between the two values 4. The cumulative density function F(x0) is the area under the probability density function f(x) from the minimum x value up to x0 x0 f(x ) = f(x)dx 0 ò xm where
    [Show full text]
  • Lecture.7 Poisson Distributions - Properties, Normal Distributions- Properties
    Lecture.7 Poisson Distributions - properties, Normal Distributions- properties Theoretical Distributions Theoretical distributions are 1. Binomial distribution Discrete distribution 2. Poisson distribution 3. Normal distribution Continuous distribution Discrete Probability distribution Bernoulli distribution A random variable x takes two values 0 and 1, with probabilities q and p ie., p(x=1) = p and p(x=0)=q, q-1-p is called a Bernoulli variate and is said to be Bernoulli distribution where p and q are probability of success and failure. It was given by Swiss mathematician James Bernoulli (1654-1705) Example • Tossing a coin(head or tail) • Germination of seed(germinate or not) Binomial distribution Binomial distribution was discovered by James Bernoulli (1654-1705). Let a random experiment be performed repeatedly and the occurrence of an event in a trial be called as success and its non-occurrence is failure. Consider a set of n independent trails (n being finite), in which the probability p of success in any trail is constant for each trial. Then q=1-p is the probability of failure in any trail. 1 The probability of x success and consequently n-x failures in n independent trails. But x successes in n trails can occur in ncx ways. Probability for each of these ways is pxqn-x. P(sss…ff…fsf…f)=p(s)p(s)….p(f)p(f)…. = p,p…q,q… = (p,p…p)(q,q…q) (x times) (n-x times) Hence the probability of x success in n trials is given by x n-x ncx p q Definition A random variable x is said to follow binomial distribution if it assumes non- negative values and its probability mass function is given by P(X=x) =p(x) = x n-x ncx p q , x=0,1,2…n q=1-p 0, otherwise The two independent constants n and p in the distribution are known as the parameters of the distribution.
    [Show full text]
  • 3.2.3 Binomial Distribution
    3.2.3 Binomial Distribution The binomial distribution is based on the idea of a Bernoulli trial. A Bernoulli trail is an experiment with two, and only two, possible outcomes. A random variable X has a Bernoulli(p) distribution if 8 > <1 with probability p X = > :0 with probability 1 − p, where 0 ≤ p ≤ 1. The value X = 1 is often termed a “success” and X = 0 is termed a “failure”. The mean and variance of a Bernoulli(p) random variable are easily seen to be EX = (1)(p) + (0)(1 − p) = p and VarX = (1 − p)2p + (0 − p)2(1 − p) = p(1 − p). In a sequence of n identical, independent Bernoulli trials, each with success probability p, define the random variables X1,...,Xn by 8 > <1 with probability p X = i > :0 with probability 1 − p. The random variable Xn Y = Xi i=1 has the binomial distribution and it the number of sucesses among n independent trials. The probability mass function of Y is µ ¶ ¡ ¢ n ¡ ¢ P Y = y = py 1 − p n−y. y For this distribution, t n EX = np, Var(X) = np(1 − p),MX (t) = [pe + (1 − p)] . 1 Theorem 3.2.2 (Binomial theorem) For any real numbers x and y and integer n ≥ 0, µ ¶ Xn n (x + y)n = xiyn−i. i i=0 If we take x = p and y = 1 − p, we get µ ¶ Xn n 1 = (p + (1 − p))n = pi(1 − p)n−i. i i=0 Example 3.2.2 (Dice probabilities) Suppose we are interested in finding the probability of obtaining at least one 6 in four rolls of a fair die.
    [Show full text]
  • 1 One Parameter Exponential Families
    1 One parameter exponential families The world of exponential families bridges the gap between the Gaussian family and general dis- tributions. Many properties of Gaussians carry through to exponential families in a fairly precise sense. • In the Gaussian world, there exact small sample distributional results (i.e. t, F , χ2). • In the exponential family world, there are approximate distributional results (i.e. deviance tests). • In the general setting, we can only appeal to asymptotics. A one-parameter exponential family, F is a one-parameter family of distributions of the form Pη(dx) = exp (η · t(x) − Λ(η)) P0(dx) for some probability measure P0. The parameter η is called the natural or canonical parameter and the function Λ is called the cumulant generating function, and is simply the normalization needed to make dPη fη(x) = (x) = exp (η · t(x) − Λ(η)) dP0 a proper probability density. The random variable t(X) is the sufficient statistic of the exponential family. Note that P0 does not have to be a distribution on R, but these are of course the simplest examples. 1.0.1 A first example: Gaussian with linear sufficient statistic Consider the standard normal distribution Z e−z2=2 P0(A) = p dz A 2π and let t(x) = x. Then, the exponential family is eη·x−x2=2 Pη(dx) / p 2π and we see that Λ(η) = η2=2: eta= np.linspace(-2,2,101) CGF= eta**2/2. plt.plot(eta, CGF) A= plt.gca() A.set_xlabel(r'$\eta$', size=20) A.set_ylabel(r'$\Lambda(\eta)$', size=20) f= plt.gcf() 1 Thus, the exponential family in this setting is the collection F = fN(η; 1) : η 2 Rg : d 1.0.2 Normal with quadratic sufficient statistic on R d As a second example, take P0 = N(0;Id×d), i.e.
    [Show full text]
  • Solutions to the Exercises
    Solutions to the Exercises Chapter 1 Solution 1.1 (a) Your computer may be programmed to allocate borderline cases to the next group down, or the next group up; and it may or may not manage to follow this rule consistently, depending on its handling of the numbers involved. Following a rule which says 'move borderline cases to the next group up', these are the five classifications. (i) 1.0-1.2 1.2-1.4 1.4-1.6 1.6-1.8 1.8-2.0 2.0-2.2 2.2-2.4 6 6 4 8 4 3 4 2.4-2.6 2.6-2.8 2.8-3.0 3.0-3.2 3.2-3.4 3.4-3.6 3.6-3.8 6 3 2 2 0 1 1 (ii) 1.0-1.3 1.3-1.6 1.6-1.9 1.9-2.2 2.2-2.5 10 6 10 5 6 2.5-2.8 2.8-3.1 3.1-3.4 3.4-3.7 7 3 1 2 (iii) 0.8-1.1 1.1-1.4 1.4-1.7 1.7-2.0 2.0-2.3 2 10 6 10 7 2.3-2.6 2.6-2.9 2.9-3.2 3.2-3.5 3.5-3.8 6 4 3 1 1 (iv) 0.85-1.15 1.15-1.45 1.45-1.75 1.75-2.05 2.05-2.35 4 9 8 9 5 2.35-2.65 2.65-2.95 2.95-3.25 3.25-3.55 3.55-3.85 7 3 3 1 1 (V) 0.9-1.2 1.2-1.5 1.5-1.8 1.8-2.1 2.1-2.4 6 7 11 7 4 2.4-2.7 2.7-3.0 3.0-3.3 3.3-3.6 3.6-3.9 7 4 2 1 1 (b) Computer graphics: the diagrams are shown in Figures 1.9 to 1.11.
    [Show full text]
  • ONE SAMPLE TESTS the Following Data Represent the Change
    1 WORKED EXAMPLES 6 INTRODUCTION TO STATISTICAL METHODS EXAMPLE 1: ONE SAMPLE TESTS The following data represent the change (in ml) in the amount of Carbon monoxide transfer (an indicator of improved lung function) in smokers with chickenpox over a one week period: 33, 2, 24, 17, 4, 1, -6 Is there evidence of significant improvement in lung function (a) if the data are normally distributed with σ = 10, (b) if the data are normally distributed with σ unknown? Use a significance level of α = 0.05. SOLUTION: (a) Here we have a sample of size 7 with sample mean x = 10.71. We want to test H0 : μ = 0.0, H1 : μ = 0.0, 6 under the assumption that the data follow a Normal distribution with σ = 10.0 known. Then, we have, in the Z-test, 10.71 0.0 z = − = 2.83, 10.0/√7 which lies in the critical region, as the critical values for this test are 1.96, for significance ± level α = 0.05. Therefore we have evidence to reject H0. The p-value is given by p = 2Φ( 2.83) = 0.004 < α. − (b) The sample variance is s2 = 14.192. In the T-test, we have test statistic t given by x 0.0 10.71 0.0 t = − = − = 2.00. s/√n 14.19/√7 The upper critical value CR is obtained by solving FSt(n 1)(CR) = 0.975, − where FSt(n 1) is the cdf of a Student-t distribution with n 1 degrees of freedom; here n = 7, so − − we can use statistical tables or a computer to find that CR = 2.447, and note that, as Student-t distributions are symmetric the lower critical value is CR.
    [Show full text]
  • Chapter 5 Sections
    Chapter 5 Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions Continuous univariate distributions: 5.6 Normal distributions 5.7 Gamma distributions Just skim 5.8 Beta distributions Multivariate distributions Just skim 5.9 Multinomial distributions 5.10 Bivariate normal distributions 1 / 43 Chapter 5 5.1 Introduction Families of distributions How: Parameter and Parameter space pf /pdf and cdf - new notation: f (xj parameters ) Mean, variance and the m.g.f. (t) Features, connections to other distributions, approximation Reasoning behind a distribution Why: Natural justification for certain experiments A model for the uncertainty in an experiment All models are wrong, but some are useful – George Box 2 / 43 Chapter 5 5.2 Bernoulli and Binomial distributions Bernoulli distributions Def: Bernoulli distributions – Bernoulli(p) A r.v. X has the Bernoulli distribution with parameter p if P(X = 1) = p and P(X = 0) = 1 − p. The pf of X is px (1 − p)1−x for x = 0; 1 f (xjp) = 0 otherwise Parameter space: p 2 [0; 1] In an experiment with only two possible outcomes, “success” and “failure”, let X = number successes. Then X ∼ Bernoulli(p) where p is the probability of success. E(X) = p, Var(X) = p(1 − p) and (t) = E(etX ) = pet + (1 − p) 8 < 0 for x < 0 The cdf is F(xjp) = 1 − p for 0 ≤ x < 1 : 1 for x ≥ 1 3 / 43 Chapter 5 5.2 Bernoulli and Binomial distributions Binomial distributions Def: Binomial distributions – Binomial(n; p) A r.v.
    [Show full text]