The Exact Joint Distribution of Conco- Mitants of Order Statistics and Their Order Statistics Under Normality

Total Page:16

File Type:pdf, Size:1020Kb

The Exact Joint Distribution of Conco- Mitants of Order Statistics and Their Order Statistics Under Normality REVSTAT – Statistical Journal Volume 11, Number 2, June 2013, 121–134 THE EXACT JOINT DISTRIBUTION OF CONCO- MITANTS OF ORDER STATISTICS AND THEIR ORDER STATISTICS UNDER NORMALITY Authors: Ayyub Sheikhi – Department of Statistics, Shahid Bahonar University of Kerman, Iran [email protected] Mahbanoo Tata – Department of Statistics, Shahid Bahonar University of Kerman, Iran [email protected] Received: December 2011 Revised: May 2012 Accepted: August 2012 Abstract: • In this work we derive the exact joint distribution of a linear combination of con- comitants of order statistics and linear combinations of their order statistics in a multivariate normal distribution. We also investigate a special case of related joint distributions discussed by He and Nagaraja (2009). Key-Words: • unified skew-normal; L-statistic; multivariate normal distribution; order statistic; concomitant; orthant probability. AMS Subject Classification: • 62G30, 62H10. 122 A. Sheikhi and M. Tata Joint Distribution of Concomitants and their Order Statistics 123 1. INTRODUCTION Suppose that the joint distribution of two n-dimensional random vectors X and Y follows a 2n dimensional multivariate normal vector with positive definite covariance matrix, i.e. X µ (1.1) ∼ N µ = x , = xx xy Y 2n µ T y Pxy Pyy ! X P P where µx, µy are respectively the mean vectors and xx, yy are the positive definite variance matrices of X and Y, while xy is their covariance matrix. Let T P P X(n) = (X1:n, X2:n,..., Xn:n) be the vector of order statistics obtained from X T P and Y(n) = (Y1:n, Y2:n,..., Yn:n) be the vector of order statistics obtained from T Y. Further, let Y[n] = (Y[1:n], Y[2:n],..., Y[n:n]) be the vector of Y -variates paired with the order statistics of X. The elements of Y[n] are called the concomitants of the order statistics of X. Nagaraja (1982) has obtained the distribution of a linear combination of order statistics from a bivariate normal random vector where the variables are exchangeable. Loperfido (2008a) has extended the results of Nagaraja (1982) to elliptical distributions. Arellano-Valle and Genton (2007) have expressed the ex- act distribution of linear combinations of order statistics from dependent random variables. Sheikhi and Jamalizadeh (2011) have showed that for arbitrary vectors T T T a and b, the distribution of (X, a Y(2), b Y(2)) is a singular skew-normal and carried out a regression analysis. Yang (1981) has studied the linear combina- tion of concomitants of order statistics. Tsukibayashi (1998) has obtained the joint distribution of (Yi:n,Y[i:n]), while He and Nagaraja (2009) have obtained the joint distribution of (Yi:n,Y[j:n]) for all i, j = 1, 2,..., n. Goel and Hall (1994) have discussed the difference between concomitants and order statistics using the n sum i=1 h(Yi:n − Y[i:n]) for some smooth function h. Recently much attention has been focused on the connection between order statistics and skew-normal P distributions (see e.g. Loperfido 2008a and 2008b and Sheikhi and Jamalizadeh T T 2011). In this article we shall obtain the joint distribution of a Y(n) and b Y[n], T T n where a = (a1, a2,..., an) and b = (b1, b2,..., bn) are arbitrary vectors in R . Since we do not assume independence, our results are more general than those of He and Nagaraja (2009). On the other hand, He and Nagaraja (2009) have not assumed normality. The concept of the skew-normal distribution was proposed independently by Roberts (1966), Ainger et al. (1977), Andel et al. (1984) and Azzalini (1985). The univariate random variable Y has a skew-normal distribution if its distribu- tion can be written as y − µ (1.2) f (y) = 2ϕ y; µ, σ2 Φ λ y ∈ R Y σ 124 A. Sheikhi and M. Tata where ϕ .; µ, σ2 is the normal density with mean µ and variance σ2 and Φ(.) denotes the standard normal distribution function. Following Arellano-Valle and Azzalini (2006), a d-dimensional random vec- tor Y is said to have a unified multivariate skew-normal distribution (Y ∼ SUNd,m (ξ, δ, Ω, Γ, Λ)), if it has a density function of the form (1.3) T −1 T −1 Φm δ + Λ Ω (y − ξ); Γ − Λ Ω Λ f (y)= ϕ (y; ξ, Ω) y ∈ Rd Y d δ Γ Φm ( ; ) where ϕd(., ξ, Ω) is the density function of a multivariate normal and Φm(.; Σ) is the multivariate normal cumulative function with the covariance matrix Σ. Γ ΛT If Σ∗ = is a singular matrix we say that the distribution of X is Λ Ω singular unified skew-normal and write SSUNd,m(ξ, δ, Ω, Γ, Λ). For more de- tails see Arellano-Valle and Azzalini (2006) and Sheikhi and Jamalizadeh (2011). In Section 2, we show that for two vectors a and b, the joint distribution of T T a Y(n) and b Y[n] belongs to the unified multivariate skew-normal family. We also discuss special cases of these distributions under the setting of independent normal random variables. Finally, in section 3 we present a numerical application of our results. 2. MAIN RESULTS Define S(X) as the class of all permutation of components of the ran- (i) dom vector X, i.e. S(X)= {X = PiX; i = 1, 2,..., n!}, where Pi is an n × n permutation matrix. Also, suppose ∆ is the difference matrix of dimension T T (n − 1) × n such that the ith row of ∆ is en, i+1− en, i, i = 1,...,n − 1, where e1, e2,..., en are n-dimensional unit basis vectors. Then ∆X = (X2 − X1, X3 − T X2,..., Xn − Xn−1) . (See e.g. Crocetta and Loperfido 2005). Further, let X(i) and Y(i) be the ith permutation of the random vectors X (i) (j) and Y respectively. We write Gij(t, ξ, )= P ∆X ≥ 0, ∆Y ≥ 0 . P ∆ Theorem 2.1. Suppose the matrix aT is of full rank. Then under bT T T T the assumption of model (1.1) the cdf of the random vector a Y(n), b Y[n] is the mixture n! n! F T T ( y , y )= F ( y , y ; ξ , δ , Γ , Ω , Λ ) G (t, ξ, ) a Y(n), b Y[n] 1 2 SUN 1 2 ij ij ij ij ij ij i j X=1 X=1 X Joint Distribution of Concomitants and their Order Statistics 125 where FSUN (. ; ξij, δij, Γij, Ωij, Λij) is the cdf of unified multivariate skew- normal with T (i) (i) (ii) T (ij) T a µy ∆µx ∆ xx ∆ ∆ xy ∆ ξ = , δij = , Γij = , ij T (i) (j) ∆ (jj) ∆T b µy ! ∆µy ! P Pyy ! P T T (ii) T (ij) (ii) (ij) a yy a a yy b ∆ xy a ∆ xy b Ωij = and Λij = bT (jj) b ∆ (ij) a ∆ (jj) b P Pyy ! Pyy Pyy ! (i) (j) where µx and µy areP respectively the mean vectors ofP the ith permutationP of the random vector X and the jth permutation of the random vector Y and (ii) (i) (jj) (j) (ij) (i) (j) xx = V ar(X ), yy = V ar(Y ) and xy = Cov(X , Y ). P P P Proof: We have T T F T T (y , y ) = P a Y ≤ y , b Y ≤ y a Y(n), b Y[n] 1 2 (n) 1 [n] 2 n! n! T (i) T (j) (ij) (ij) = P a Y ≤ y1, b Y ≤ y2|A P A i j X=1 X=1 where A(ij) = {∆X(i) ≥ 0, ∆Y(j) ≥ 0}. Since ∆X(i) ∆ 0 0 X(i) ∆Y(j) 0 ∆ 0 (2.1) = Y(j) , aT Y(i) 0 0aT Y(i) bT Y(j) 0 bT 0 3n×1 2n×1 2n×3n the full rank assumption implies nonsingularity of the matrix on the right hand side of (2.1). Furthermore, ∆X(i) ∆Y(j) aT Y(i) bT Y(j) (i) (ii) T (ij) T (ii) (ij) ∆µx ∆ xx ∆ ∆ xy ∆ ∆ xy a ∆ xy b ∆µ(j) ∆ (jj) ∆T ∆ (ij) a ∆ (jj) b ∼ N y , P Pyy Pyy Pyy . 2n T (i) T (ii) T (ij) a µy a a a b P P yy P yy bT µ(j) bT (jj) b y P Pyy P Now, similar to Sheikhi and Jamalizadeh (2011), we immediately conclude that T T (i) T (j) (i) (j) a Y , b Y | ∆X ≥ 0, ∆Y ≥ 0 ∼ SUN2, 2(n−1)(ξij, δij, Γij, Ωij, Λij) . This proves the Theorem. 126 A. Sheikhi and M. Tata T Remark 2.1. If the rank of the matrix ∆, aT , bT is at most n − 1, the T T T joint distribution of a Y(n), b Y[n] is a mixture of a unified skew-normals and a singular unified skew-normals. In this section we assume that the matrix T ∆, aT , bT is of full rank. A special case will be investigated later in the paper. Let (Xi,Yi), i = 1, 2,..., n be a random sample of size n from a bivariate 2 2 normal N2 µx, µy, σx, σy, ρ , then the model (1.1) reduces to the following: X µx1n xx xy (2.2) ∼ N2n µ = , = Y µy1n yy X P P 2 2 T where xx = σxIn, yy = σyIn and xy = ρσxσy1n1n whereP ρ is the correla- tion coefficient between X and Y. P P P The following corollary describes the joint distribution of a linear combi- nation of concomitants of order statistics and a linear combination of their order statistics under the independence assumption.
Recommended publications
  • On the Meaning and Use of Kurtosis
    Psychological Methods Copyright 1997 by the American Psychological Association, Inc. 1997, Vol. 2, No. 3,292-307 1082-989X/97/$3.00 On the Meaning and Use of Kurtosis Lawrence T. DeCarlo Fordham University For symmetric unimodal distributions, positive kurtosis indicates heavy tails and peakedness relative to the normal distribution, whereas negative kurtosis indicates light tails and flatness. Many textbooks, however, describe or illustrate kurtosis incompletely or incorrectly. In this article, kurtosis is illustrated with well-known distributions, and aspects of its interpretation and misinterpretation are discussed. The role of kurtosis in testing univariate and multivariate normality; as a measure of departures from normality; in issues of robustness, outliers, and bimodality; in generalized tests and estimators, as well as limitations of and alternatives to the kurtosis measure [32, are discussed. It is typically noted in introductory statistics standard deviation. The normal distribution has a kur- courses that distributions can be characterized in tosis of 3, and 132 - 3 is often used so that the refer- terms of central tendency, variability, and shape. With ence normal distribution has a kurtosis of zero (132 - respect to shape, virtually every textbook defines and 3 is sometimes denoted as Y2)- A sample counterpart illustrates skewness. On the other hand, another as- to 132 can be obtained by replacing the population pect of shape, which is kurtosis, is either not discussed moments with the sample moments, which gives or, worse yet, is often described or illustrated incor- rectly. Kurtosis is also frequently not reported in re- ~(X i -- S)4/n search articles, in spite of the fact that virtually every b2 (•(X i - ~')2/n)2' statistical package provides a measure of kurtosis.
    [Show full text]
  • The Probability Lifesaver: Order Statistics and the Median Theorem
    The Probability Lifesaver: Order Statistics and the Median Theorem Steven J. Miller December 30, 2015 Contents 1 Order Statistics and the Median Theorem 3 1.1 Definition of the Median 5 1.2 Order Statistics 10 1.3 Examples of Order Statistics 15 1.4 TheSampleDistributionoftheMedian 17 1.5 TechnicalboundsforproofofMedianTheorem 20 1.6 TheMedianofNormalRandomVariables 22 2 • Greetings again! In this supplemental chapter we develop the theory of order statistics in order to prove The Median Theorem. This is a beautiful result in its own, but also extremely important as a substitute for the Central Limit Theorem, and allows us to say non- trivial things when the CLT is unavailable. Chapter 1 Order Statistics and the Median Theorem The Central Limit Theorem is one of the gems of probability. It’s easy to use and its hypotheses are satisfied in a wealth of problems. Many courses build towards a proof of this beautiful and powerful result, as it truly is ‘central’ to the entire subject. Not to detract from the majesty of this wonderful result, however, what happens in those instances where it’s unavailable? For example, one of the key assumptions that must be met is that our random variables need to have finite higher moments, or at the very least a finite variance. What if we were to consider sums of Cauchy random variables? Is there anything we can say? This is not just a question of theoretical interest, of mathematicians generalizing for the sake of generalization. The following example from economics highlights why this chapter is more than just of theoretical interest.
    [Show full text]
  • Theoretical Statistics. Lecture 5. Peter Bartlett
    Theoretical Statistics. Lecture 5. Peter Bartlett 1. U-statistics. 1 Outline of today’s lecture We’ll look at U-statistics, a family of estimators that includes many interesting examples. We’ll study their properties: unbiased, lower variance, concentration (via an application of the bounded differences inequality), asymptotic variance, asymptotic distribution. (See Chapter 12 of van der Vaart.) First, we’ll consider the standard unbiased estimate of variance—a special case of a U-statistic. 2 Variance estimates n 1 s2 = (X − X¯ )2 n n − 1 i n Xi=1 n n 1 = (X − X¯ )2 +(X − X¯ )2 2n(n − 1) i n j n Xi=1 Xj=1 n n 1 2 = (X − X¯ ) − (X − X¯ ) 2n(n − 1) i n j n Xi=1 Xj=1 n n 1 1 = (X − X )2 n(n − 1) 2 i j Xi=1 Xj=1 1 1 = (X − X )2 . n 2 i j 2 Xi<j 3 Variance estimates This is unbiased for i.i.d. data: 1 Es2 = E (X − X )2 n 2 1 2 1 = E ((X − EX ) − (X − EX ))2 2 1 1 2 2 1 2 2 = E (X1 − EX1) +(X2 − EX2) 2 2 = E (X1 − EX1) . 4 U-statistics Definition: A U-statistic of order r with kernel h is 1 U = n h(Xi1 ,...,Xir ), r iX⊆[n] where h is symmetric in its arguments. [If h is not symmetric in its arguments, we can also average over permutations.] “U” for “unbiased.” Introduced by Wassily Hoeffding in the 1940s. 5 U-statistics Theorem: [Halmos] θ (parameter, i.e., function defined on a family of distributions) admits an unbiased estimator (ie: for all sufficiently large n, some function of the i.i.d.
    [Show full text]
  • Lecture 14 Testing for Kurtosis
    9/8/2016 CHE384, From Data to Decisions: Measurement, Kurtosis Uncertainty, Analysis, and Modeling • For any distribution, the kurtosis (sometimes Lecture 14 called the excess kurtosis) is defined as Testing for Kurtosis 3 (old notation = ) • For a unimodal, symmetric distribution, Chris A. Mack – a positive kurtosis means “heavy tails” and a more Adjunct Associate Professor peaked center compared to a normal distribution – a negative kurtosis means “light tails” and a more spread center compared to a normal distribution http://www.lithoguru.com/scientist/statistics/ © Chris Mack, 2016Data to Decisions 1 © Chris Mack, 2016Data to Decisions 2 Kurtosis Examples One Impact of Excess Kurtosis • For the Student’s t • For a normal distribution, the sample distribution, the variance will have an expected value of s2, excess kurtosis is and a variance of 6 2 4 1 for DF > 4 ( for DF ≤ 4 the kurtosis is infinite) • For a distribution with excess kurtosis • For a uniform 2 1 1 distribution, 1 2 © Chris Mack, 2016Data to Decisions 3 © Chris Mack, 2016Data to Decisions 4 Sample Kurtosis Sample Kurtosis • For a sample of size n, the sample kurtosis is • An unbiased estimator of the sample excess 1 kurtosis is ∑ ̅ 1 3 3 1 6 1 2 3 ∑ ̅ Standard Error: • For large n, the sampling distribution of 1 24 2 1 approaches Normal with mean 0 and variance 2 1 of 24/n 3 5 • For small samples, this estimator is biased D. N. Joanes and C. A. Gill, “Comparing Measures of Sample Skewness and Kurtosis”, The Statistician, 47(1),183–189 (1998).
    [Show full text]
  • Covariances of Two Sample Rank Sum Statistics
    JOURNAL OF RESEARCH of the National Bureou of Standards - B. Mathematical Sciences Volume 76B, Nos. 1 and 2, January-June 1972 Covariances of Two Sample Rank Sum Statistics Peter V. Tryon Institute for Basic Standards, National Bureau of Standards, Boulder, Colorado 80302 (November 26, 1971) This note presents an elementary derivation of the covariances of the e(e - 1)/2 two-sample rank sum statistics computed among aU pairs of samples from e populations. Key words: e Sample proble m; covariances, Mann-Whitney-Wilcoxon statistics; rank sum statistics; statistics. Mann-Whitney or Wilcoxon rank sum statistics, computed for some or all of the c(c - 1)/2 pairs of samples from c populations, have been used in testing the null hypothesis of homogeneity of dis­ tribution against a variety of alternatives [1, 3,4,5).1 This note presents an elementary derivation of the covariances of such statistics under the null hypothesis_ The usual approach to such an analysis is the rank sum viewpoint of the Wilcoxon form of the statistic. Using this approach, Steel [3] presents a lengthy derivation of the covariances. In this note it is shown that thinking in terms of the Mann-Whitney form of the statistic leads to an elementary derivation. For comparison and completeness the rank sum derivation of Kruskal and Wallis [2] is repeated in obtaining the means and variances. Let x{, r= 1,2, ... , ni, i= 1,2, ... , c, be the rth item in the sample of size ni from the ith of c populations. Let Mij be the Mann-Whitney statistic between the ith andjth samples defined by n· n· Mij= ~ t z[J (1) s=1 1' = 1 where l,xJ~>xr Zr~= { I) O,xj";;;x[ } Thus Mij is the number of times items in the jth sample exceed items in the ith sample.
    [Show full text]
  • Analysis of Covariance (ANCOVA) with Two Groups
    NCSS Statistical Software NCSS.com Chapter 226 Analysis of Covariance (ANCOVA) with Two Groups Introduction This procedure performs analysis of covariance (ANCOVA) for a grouping variable with 2 groups and one covariate variable. This procedure uses multiple regression techniques to estimate model parameters and compute least squares means. This procedure also provides standard error estimates for least squares means and their differences, and computes the T-test for the difference between group means adjusted for the covariate. The procedure also provides response vs covariate by group scatter plots and residuals for checking model assumptions. This procedure will output results for a simple two-sample equal-variance T-test if no covariate is entered and simple linear regression if no group variable is entered. This allows you to complete the ANCOVA analysis if either the group variable or covariate is determined to be non-significant. For additional options related to the T- test and simple linear regression analyses, we suggest you use the corresponding procedures in NCSS. The group variable in this procedure is restricted to two groups. If you want to perform ANCOVA with a group variable that has three or more groups, use the One-Way Analysis of Covariance (ANCOVA) procedure. This procedure cannot be used to analyze models that include more than one covariate variable or more than one group variable. If the model you want to analyze includes more than one covariate variable and/or more than one group variable, use the General Linear Models (GLM) for Fixed Factors procedure instead. Kinds of Research Questions A large amount of research consists of studying the influence of a set of independent variables on a response (dependent) variable.
    [Show full text]
  • What Is Statistic?
    What is Statistic? OPRE 6301 In today’s world. ...we are constantly being bombarded with statistics and statistical information. For example: Customer Surveys Medical News Demographics Political Polls Economic Predictions Marketing Information Sales Forecasts Stock Market Projections Consumer Price Index Sports Statistics How can we make sense out of all this data? How do we differentiate valid from flawed claims? 1 What is Statistics?! “Statistics is a way to get information from data.” Statistics Data Information Data: Facts, especially Information: Knowledge numerical facts, collected communicated concerning together for reference or some particular fact. information. Statistics is a tool for creating an understanding from a set of numbers. Humorous Definitions: The Science of drawing a precise line between an unwar- ranted assumption and a forgone conclusion. The Science of stating precisely what you don’t know. 2 An Example: Stats Anxiety. A business school student is anxious about their statistics course, since they’ve heard the course is difficult. The professor provides last term’s final exam marks to the student. What can be discerned from this list of numbers? Statistics Data Information List of last term’s marks. New information about the statistics class. 95 89 70 E.g. Class average, 65 Proportion of class receiving A’s 78 Most frequent mark, 57 Marks distribution, etc. : 3 Key Statistical Concepts. Population — a population is the group of all items of interest to a statistics practitioner. — frequently very large; sometimes infinite. E.g. All 5 million Florida voters (per Example 12.5). Sample — A sample is a set of data drawn from the population.
    [Show full text]
  • Lecture 14: Order Statistics; Conditional Expectation
    Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 14: Order Statistics; Conditional Expectation Relevant textbook passages: Pitman [5]: Section 4.6 Larsen–Marx [2]: Section 3.11 14.1 Order statistics Pitman [5]: Given a random vector (X(1,...,Xn) on the probability) space (S, E,P ), for each s ∈ S, sort the § 4.6 components into a vector X(1)(s),...,X(n)(s) satisfying X(1)(s) 6 X(2)(s) 6 ··· 6 X(n)(s). The vector (X(1),...,X(n)) is called the vector of order statistics of (X1,...,Xn). Equivalently, { } X(k) = min max{Xj : j ∈ J} : J ⊂ {1, . , n} & |J| = k . Order statistics play an important role in the study of auctions, among other things. 14.2 Marginal Distribution of Order Statistics From here on, we shall assume that the original random variables X1,...,Xn are independent and identically distributed. Let X1,...,Xn be independent and identically distributed random variables with common cumulative distribution function F ,( and let (X) (1),...,X(n)) be the vector of order statistics of X1,...,Xn. By breaking the event X(k) 6 x into simple disjoint subevents, we get ( ) X(k) 6 x = ( ) X(n) 6 x ∪ ( ) X(n) > x, X(n−1) 6 x . ∪ ( ) X(n) > x, . , X(j+1) > x, X(j) 6 x . ∪ ( ) X(n) > x, . , X(k+1) > x, X(k) 6 x . 14–1 Ma 3/103 Winter 2017 KC Border Order Statistics; Conditional Expectation 14–2 Each of these subevents is disjoint from the ones above it, and each has a binomial probability: ( ) X(n) > x, .
    [Show full text]
  • Medians and Order Statistics
    Medians and Order Statistics CLRS Chapter 9 1 What Are Order Statistics? The k-th order statistic is the k-th smallest element of an array. 3 4 13 14 23 27 41 54 65 75 8th order statistic ⎢n⎥ ⎢ ⎥ The lower median is the ⎣ 2 ⎦ -th order statistic ⎡n⎤ ⎢ ⎥ The upper median is the ⎢ 2 ⎥ -th order statistic If n is odd, lower and upper median are the same 3 4 13 14 23 27 41 54 65 75 lower median upper median What are Order Statistics? Selecting ith-ranked item from a collection. – First: i = 1 – Last: i = n ⎢n⎥ ⎡n⎤ – Median(s): i = ⎢ ⎥,⎢ ⎥ ⎣2⎦ ⎢2⎥ 3 Order Statistics Overview • Assume collection is unordered, otherwise trivial. find ith order stat = A[i] • Can sort first – Θ(n lg n), but can do better – Θ(n). • I can find max and min in Θ(n) time (obvious) • Can we find any order statistic in linear time? (not obvious!) 4 Order Statistics Overview How can we modify Quicksort to obtain expected-case Θ(n)? ? ? Pivot, partition, but recur only on one set of data. No join. 5 Using the Pivot Idea • Randomized-Select(A[p..r],i) looking for ith o.s. if p = r return A[p] q <- Randomized-Partition(A,p,r) k <- q-p+1 the size of the left partition if i=k then the pivot value is the answer return A[q] else if i < k then the answer is in the front return Randomized-Select(A,p,q-1,i) else then the answer is in the back half return Randomized-Select(A,q+1,r,i-k) 6 Randomized Selection • Analyzing RandomizedSelect() – Worst case: partition always 0:n-1 T(n) = T(n-1) + O(n) = O(n2) • No better than sorting! – “Best” case: suppose a 9:1 partition T(n) = T(9n/10)
    [Show full text]
  • Exponential Families and Theoretical Inference
    EXPONENTIAL FAMILIES AND THEORETICAL INFERENCE Bent Jørgensen Rodrigo Labouriau August, 2012 ii Contents Preface vii Preface to the Portuguese edition ix 1 Exponential families 1 1.1 Definitions . 1 1.2 Analytical properties of the Laplace transform . 11 1.3 Estimation in regular exponential families . 14 1.4 Marginal and conditional distributions . 17 1.5 Parametrizations . 20 1.6 The multivariate normal distribution . 22 1.7 Asymptotic theory . 23 1.7.1 Estimation . 25 1.7.2 Hypothesis testing . 30 1.8 Problems . 36 2 Sufficiency and ancillarity 47 2.1 Sufficiency . 47 2.1.1 Three lemmas . 48 2.1.2 Definitions . 49 2.1.3 The case of equivalent measures . 50 2.1.4 The general case . 53 2.1.5 Completeness . 56 2.1.6 A result on separable σ-algebras . 59 2.1.7 Sufficiency of the likelihood function . 60 2.1.8 Sufficiency and exponential families . 62 2.2 Ancillarity . 63 2.2.1 Definitions . 63 2.2.2 Basu's Theorem . 65 2.3 First-order ancillarity . 67 2.3.1 Examples . 67 2.3.2 Main results . 69 iii iv CONTENTS 2.4 Problems . 71 3 Inferential separation 77 3.1 Introduction . 77 3.1.1 S-ancillarity . 81 3.1.2 The nonformation principle . 83 3.1.3 Discussion . 86 3.2 S-nonformation . 91 3.2.1 Definition . 91 3.2.2 S-nonformation in exponential families . 96 3.3 G-nonformation . 99 3.3.1 Transformation models . 99 3.3.2 Definition of G-nonformation . 103 3.3.3 Cox's proportional risks model .
    [Show full text]
  • Lecture 8 — October 15 8.1 Bayes Estimators and Average Risk
    STATS 300A: Theory of Statistics Fall 2015 Lecture 8 | October 15 Lecturer: Lester Mackey Scribe: Hongseok Namkoong, Phan Minh Nguyen Warning: These notes may contain factual and/or typographic errors. 8.1 Bayes Estimators and Average Risk Optimality 8.1.1 Setting We discuss the average risk optimality of estimators within the framework of Bayesian de- cision problems. As with the general decision problem setting the Bayesian setup considers a model P = fPθ : θ 2 Ωg, for our data X, a loss function L(θ; d), and risk R(θ; δ). In the frequentist approach, the parameter θ was considered to be an unknown deterministic quan- tity. In the Bayesian paradigm, we consider a measure Λ over the parameter space which we call a prior. Assuming this measure defines a probability distribution, we interpret the parameter θ as an outcome of the random variable Θ ∼ Λ. So, in this setup both X and θ are random. Conditioning on Θ = θ, we assume the data is generated by the distribution Pθ. Now, the optimality goal for our decision problem of estimating g(θ) is the minimization of the average risk r(Λ; δ) = E[L(Θ; δ(X))] = E[E[L(Θ; δ(X)) j X]]: An estimator δ which minimizes this average risk is a Bayes estimator and is sometimes referred to as being Bayes. Note that the average risk is an expectation over both the random variables Θ and X. Then by using the tower property, we showed last time that it suffices to find an estimator δ which minimizes the posterior risk E[L(Θ; δ(X))jX = x] for almost every x.
    [Show full text]
  • Decomposing High-Order Statistics for Sensitivity Analysis
    Center for Turbulence Research 139 Annual Research Briefs 2014 Decomposing high-order statistics for sensitivity analysis By G. Geraci, P.M. Congedo† AND G. Iaccarino 1. Motivation and objectives Sensitivity analysis in the presence of uncertainties in operating conditions, material properties, and manufacturing tolerances poses a tremendous challenge to the scientific computing community. In particular, in realistic situations, the presence of a large number of uncertain inputs complicates the task of propagation and assessment of output un- certainties; many of the popular techniques, such as stochastic collocation or polynomial chaos, lead to exponentially increasing costs, thus making these methodologies unfeasible (Foo & Karniadakis 2010). Handling uncertain parameters becomes even more challeng- ing when robust design optimization is of interest (Kim et al. 2006; Eldred 2009). One of the alternative solutions for reducing the cost of the Uncertainty Quantification (UQ) methods is based on approaches attempting to identify the relative importance of the in- put uncertainties. In the literature, global sensitivity analysis (GSA) aims at quantifying how uncertainties in the input parameters of a model contribute to the uncertainties in its output (Borgonovo et al. 2003). Traditionally, GSA is performed using methods based on the decomposition of the output variance (Sobol 2001), i.e., ANalysis Of VAriance, ANOVA. The ANOVA approach involves splitting a multi-dimensional function into its contributions from different groups of dependent variables. The ANOVA-based analysis creates a hierarchy of dominant input parameters, for a given output, when variations are computed in terms of variance. A limitation of this approach is the fact that it is based on the variance since it might not be a sufficient indicator of the overall output variations.
    [Show full text]