A Kurtosis-Based Dynamic Approach to Gaussian Mixture Modeling

Total Page:16

File Type:pdf, Size:1020Kb

A Kurtosis-Based Dynamic Approach to Gaussian Mixture Modeling IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART A: SYSTEMS AND HUMANS, VOL. 29, NO. 4, JULY 1999 393 ACKNOWLEDGMENT [25] L. Staiger, “Quadtrees and the Hausdorff dimension of pictures,” in Workshop Geometrical Problems Image Processing, Georgental, GDR, The author thanks R. Riedi, L. Barreira, and L. Olsen for helpful 1989, pp. 173–178. comments on the multifractal formalism. He is also thankful to [26] K. Culik, II and S. Dube, “Affine automata and related techniques for generation of complex images,” Theor. Comput. Sci., vol. 116, no. 2, anonymous reviewers and G. Dorffner for helpful comments on the pp. 373–398, 1993. manuscript. [27] , “Rational and affine expressions for image description,” Discrete Appl. Math., vol. 41, pp. 85–120, 1993. [28] Y. Pesin and H. Weiss, “On the dimension of deterministic and cantor- like sets, symbolic dynamics, and the Eckmann–Ruelle conjecture,” REFERENCES Commun. Math. Phys, vol. 182, no. 1, pp. 105–153, 1996. [29] P. Moran, “Additive functions of intervals and Hausdorff dimension,” [1] G. Mayer-Kress, R. Bargar, and I. Choi, “Musical structures in data in Proc. Cambridge Philos. Soc., 1946, vol. 42, pp. 15–23. from chaotic attractors,” Tech. Rep. CCSR-92-14, Center for Complex [30] Y. Pesin and H. Weiss, “A multifractal analysis of equilibrium measures Syst. Res., Beckman Inst., Univ. Illinois, Urbana-Champaign, 1992. for conformal expanding maps and Moran-like geometric constructions,” [2] C. L. Berthelsen, J. A. Glazier, and M. H. Skolnik, “Global fractal J. Stat. Phys., vol. 86, no. 1/2, pp. 233–275, 1997. dimension of human DNA sequences treated as pseudorandom walks,” [31] L. Barreira, Y. Pesin, and J. Schmeling, “On a general concept of multi- Phys. Rev. A, vol. 45, no. 12, pp. 8902–8913, 1992. fractality: Multifractal spectra for dimensions, entropies, and Lyapunov [3] H. E. Stanley, S. V. Buldyrev, A. L. Goldberger, Z. D. Goldberger, exponents multifractal rigidity,” Chaos: Interdiscipl. J. Nonlinear Sci., S. Havlin, R. N. Mantegna, S. M. Ossadnik, C.-K. Peng, and M. vol. 7, no. 1, pp. 27–53, 1996. Simons, “Statistical mechanics in biology: How ubiquitous are long- [32] D. Ron, Y. Singer, and N. Tishby, “The power of amnesia,” Mach. range correlations?,” Phys. A, vol. 205, pp. 214–253, 1994. Learn., vol. 25, 1996. [4] J. Jeffrey, “Chaos game representation of gene structure,” Nucl. Acids [33] M. J. Weinberger, J. J. Rissanen, and M. Feder, “A universal finite Res., vol. 18, no. 8, pp. 2163–2170, 1990. memory source,” IEEE Trans. Inform. Theory, vol. 41, pp. 643–652, [5] M. F. Barnsley, Fractals Everywhere. New York: Academic, 1988. May 1995. [6] J. L. Oliver, P. Bernaola-Galvan,´ J. Guerrero-Garcia, and R. Roman´ [34] P. Tinoˇ and G. Dorffner, “Constructing finite-context sources from Roldan, “Entropic profiles of DNA sequences through chaos-game- fractal representations of symbolic sequences,” Tech. Rep. TR-98-18, derived images,” J. Theor. Biol., no. 160, pp. 457–470, 1993. Austrian Res. Inst. Artif. Intell., 1998. [7] R. Roman-Roldan, P. Bernaola-Galvan, and J. L. Oliver, “Entropic feature for sequence pattern through iteration function systems,” Pattern Recognit. Lett., vol. 15, pp. 567–573, 1994. [8] S. Basu, A. Pan, C. Dutta, and D. J, “Chaos game representation of proteins,” J. Mol. Graph. Model., vol. 15, no. 5, pp. 279–289, 1997. [9] V. V. Solovyev, S. V. Korolev, and H. A. Lim, “A new approach for the classification of functional regions of DNA sequences based on fractal A Kurtosis-Based Dynamic Approach representation,” Int. J. Genom. Res., vol. 1, no. 1, pp. 109–128, 1993. to Gaussian Mixture Modeling [10] A. Fiser, G. E. Tusnady, and I. Simon, “Chaos game representation of protein structures,” J. Mol. Graph., vol. 12, no. 4, pp. 302–304, 1994. Nikos Vlassis and Aristidis Likas [11] K. A. Hill and S. M. Singh, “The evolution of species-type specificity in the global DNA sequence organization of mitochondrial genomes,” Genome, vol. 40, pp. 342–356, 1997. [12] W. Li, “The study of correlation structures of DNA sequences: A critical Abstract— We address the problem of probability density function review,” Comput. Chem., vol. 21, no. 4, pp. 257–272, 1997. estimation using a Gaussian mixture model updated with the expectation- [13] R. Roman-Roldan, P. Bernaola-Galvan, and J. L. Oliver, “Application maximization (EM) algorithm. To deal with the case of an unknown of information theory to DNA sequence analysis: A review,” Pattern number of mixing kernels, we define a new measure for Gaussian Recognit., vol. 29, no. 7, pp. 1187–1194, 1996. mixtures, called total kurtosis, which is based on the weighted sample [14] Y. Pesin, Dimension Theory in Dynamical Systems: Rigorous Results kurtoses of the kernels. This measure provides an indication of how well and Applications. Chicago, IL: Univ. of Chicago Press, 1997. the Gaussian mixture fits the data. Then we propose a new dynamic [15] P. Tinoˇ and M. Kotele¨ s,˘ “Extracting finite state representations from algorithm for Gaussian mixture density estimation which monitors the recurrent neural networks trained on chaotic symbolic sequences,” IEEE total kurtosis at each step of the EM algorithm in order to decide Trans. Neural Networks, vol. 10, pp. 284–302, Mar. 1999. dynamically on the correct number of kernels and possibly escape from [16] R. Kenyon and Y. Peres, “Measures of full dimension on affine invariant local maxima. We show the potential of our technique in approximating sets,” Ergod. Theory Dynam. Syst., vol. 16, pp. 307–323, 1996. unknown densities through a series of examples with several density [17] A. I. Khinchin, Mathematical Foundations of Information Theory. New estimation problems. York: Dover, 1957. [18] A. Renyi, “On the dimension and entropy of probability distributions,” Index Terms— Expectation-maximization (EM) algorithm, Gaussian Acta Math. Hung., no. 10, pp. 193, 1959. mixture modeling, number of mixing kernels, probability density function [19] P. Grassberger, “Information and complexity measures in dynamical sys- estimation, total kurtosis, weighted kurtosis. tems,” in Information Dynamics, H. Atmanspacher and H. Scheingraber, Eds. New York: Plenum, 1991, pp. 15–33. [20] J. P. Crutchfield and K. Young, “Computation at the onset of chaos,” I. INTRODUCTION in Complexity, Entropy, and the Physics of Information, SFI Studies in The Gaussian mixture model [1] has been proposed as a general the Sciences of Complexity, W. H. Zurek, Ed. Reading, MA: Addison- Wesley, 1990, pp. 223–269, vol. 8. model for estimating an unknown probability density function, or [21] C. Beck and F. Schlogl, Thermodynamics of Chaotic Systems. Cam- simply density. The virtues of the model lie mainly in its good bridge, U.K.: Cambridge Univ. Press, 1995. [22] J. L. McCauley, Chaos, Dynamics and Fractals: An Algorithmic Ap- Manuscript received September 17, 1998. proach to Deterministic Chaos. Cambridge, U.K.: Cambridge Univ. N. Vlassis is with the RWCP, Autonomous Learning Functions SNN, Press, 1994. Department of Computer Science, University of Amsterdam, 1098 SJ Am- [23] R. H. Riedi, “Conditional and relative multifractal spectra,” Fractals, sterdam, The Netherlands (e-mail: [email protected]). vol. 5, no. 1, pp. 153–168, 1997. A. Likas is with the Department of Computer Science, University of [24] K. J. Falconer, Fractal Geometry: Mathematical Foundations and Ap- Ioannina, 45110 Ioannina, Greece (e-mail: [email protected]). plications. New York: Wiley, 1990. Publisher Item Identifier S 1083-4427(99)05344-8. 1083–4427/99$10.00 1999 IEEE 394 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART A: SYSTEMS AND HUMANS, VOL. 29, NO. 4, JULY 1999 approximation properties and the variety of estimation algorithms that kurtosis that is needed by the algorithm and then show how the algo- exist in the literature [1], [2]. The model assumes that the unknown rithm can make use of this quantity to produce better solutions. We density can be written as a weighted finite sum of Gaussian kernels, consider the univariate case only. Work is under progress to extend the with different mixing weights and different parameters, namely, definition and use of total kurtosis in higher dimensions. Section IV means and covariance matrices. Then, depending on the estimation gives experimental results from the application of the algorithm to algorithm, an optimum vector of these parameters is sought that density estimation problems, while Section V summarizes and gives optimizes some criterion. Most often, the estimation of the parameters hints for future research. of the mixture is carried out by the maximum likelihood method, aiming at maximizing the likelihood of a set of samples drawn II. GAUSSIAN MIXTURES AND THE EM ALGORITHM independently from the unknown density. One of the algorithms often used for Gaussian mixture model- A. Gaussian Mixtures ing is the expectation-maximization (EM) algorithm, a well-known x statistical tool for maximum likelihood problems [3]. The algorithm We say that a random variable has a finite mixture distribution p@xA provides iterative formulae for the estimation of the unknown param- when its probability density function can be written as a finite eters of the mixture, and can be proven to monotonically increase in weighted sum of known densities, or simply kernels. In cases where x each step the likelihood function. However, the main drawbacks of each kernel is the Gaussian density, we say that follows a Gaussian u EM is that it requires an initialization of the parameter vector near the mixture. For the univariate case and for a number of Gaussian solution, and also it assumes that the total number of mixing kernels kernels, the unknown mixture can be written is known in advance. u To overcome the above limitations, we propose in this paper a p@xAa %j p@xjjA (1) novel dynamic algorithm for Gaussian mixture modeling that starts jaI u u aI with a small number of kernels (usually ) and performs EM where p@xjjA stands for the univariate Gaussian x@"j Y'j A steps in order to maximize the likelihood of the data, while at the same I @x " AP time monitors the value of a new measure of the mixture, called total p@xjjAa p exp j P'P (2) kurtosis, that indicates how well the Gaussian mixture fits the input 'j P% j data.
Recommended publications
  • A Tail Quantile Approximation Formula for the Student T and the Symmetric Generalized Hyperbolic Distribution
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Schlüter, Stephan; Fischer, Matthias J. Working Paper A tail quantile approximation formula for the student t and the symmetric generalized hyperbolic distribution IWQW Discussion Papers, No. 05/2009 Provided in Cooperation with: Friedrich-Alexander University Erlangen-Nuremberg, Institute for Economics Suggested Citation: Schlüter, Stephan; Fischer, Matthias J. (2009) : A tail quantile approximation formula for the student t and the symmetric generalized hyperbolic distribution, IWQW Discussion Papers, No. 05/2009, Friedrich-Alexander-Universität Erlangen-Nürnberg, Institut für Wirtschaftspolitik und Quantitative Wirtschaftsforschung (IWQW), Nürnberg This Version is available at: http://hdl.handle.net/10419/29554 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten, If the documents have been made available under an Open gelten abweichend von diesen Nutzungsbedingungen die in der dort Content Licence (especially Creative Commons Licences), you genannten Lizenz gewährten Nutzungsrechte. may exercise further usage rights as specified in the indicated licence. www.econstor.eu IWQW Institut für Wirtschaftspolitik und Quantitative Wirtschaftsforschung Diskussionspapier Discussion Papers No.
    [Show full text]
  • Mixture Random Effect Model Based Meta-Analysis for Medical Data
    Mixture Random Effect Model Based Meta-analysis For Medical Data Mining Yinglong Xia?, Shifeng Weng?, Changshui Zhang??, and Shao Li State Key Laboratory of Intelligent Technology and Systems,Department of Automation, Tsinghua University, Beijing, China [email protected], [email protected], fzcs,[email protected] Abstract. As a powerful tool for summarizing the distributed medi- cal information, Meta-analysis has played an important role in medical research in the past decades. In this paper, a more general statistical model for meta-analysis is proposed to integrate heterogeneous medi- cal researches efficiently. The novel model, named mixture random effect model (MREM), is constructed by Gaussian Mixture Model (GMM) and unifies the existing fixed effect model and random effect model. The pa- rameters of the proposed model are estimated by Markov Chain Monte Carlo (MCMC) method. Not only can MREM discover underlying struc- ture and intrinsic heterogeneity of meta datasets, but also can imply reasonable subgroup division. These merits embody the significance of our methods for heterogeneity assessment. Both simulation results and experiments on real medical datasets demonstrate the performance of the proposed model. 1 Introduction As the great improvement of experimental technologies, the growth of the vol- ume of scientific data relevant to medical experiment researches is getting more and more massively. However, often the results spreading over journals and on- line database appear inconsistent or even contradict because of variance of the studies. It makes the evaluation of those studies to be difficult. Meta-analysis is statistical technique for assembling to integrate the findings of a large collection of analysis results from individual studies.
    [Show full text]
  • A Family of Skew-Normal Distributions for Modeling Proportions and Rates with Zeros/Ones Excess
    S S symmetry Article A Family of Skew-Normal Distributions for Modeling Proportions and Rates with Zeros/Ones Excess Guillermo Martínez-Flórez 1, Víctor Leiva 2,* , Emilio Gómez-Déniz 3 and Carolina Marchant 4 1 Departamento de Matemáticas y Estadística, Facultad de Ciencias Básicas, Universidad de Córdoba, Montería 14014, Colombia; [email protected] 2 Escuela de Ingeniería Industrial, Pontificia Universidad Católica de Valparaíso, 2362807 Valparaíso, Chile 3 Facultad de Economía, Empresa y Turismo, Universidad de Las Palmas de Gran Canaria and TIDES Institute, 35001 Canarias, Spain; [email protected] 4 Facultad de Ciencias Básicas, Universidad Católica del Maule, 3466706 Talca, Chile; [email protected] * Correspondence: [email protected] or [email protected] Received: 30 June 2020; Accepted: 19 August 2020; Published: 1 September 2020 Abstract: In this paper, we consider skew-normal distributions for constructing new a distribution which allows us to model proportions and rates with zero/one inflation as an alternative to the inflated beta distributions. The new distribution is a mixture between a Bernoulli distribution for explaining the zero/one excess and a censored skew-normal distribution for the continuous variable. The maximum likelihood method is used for parameter estimation. Observed and expected Fisher information matrices are derived to conduct likelihood-based inference in this new type skew-normal distribution. Given the flexibility of the new distributions, we are able to show, in real data scenarios, the good performance of our proposal. Keywords: beta distribution; centered skew-normal distribution; maximum-likelihood methods; Monte Carlo simulations; proportions; R software; rates; zero/one inflated data 1.
    [Show full text]
  • D. Normal Mixture Models and Elliptical Models
    D. Normal Mixture Models and Elliptical Models 1. Normal Variance Mixtures 2. Normal Mean-Variance Mixtures 3. Spherical Distributions 4. Elliptical Distributions QRM 2010 74 D1. Multivariate Normal Mixture Distributions Pros of Multivariate Normal Distribution • inference is \well known" and estimation is \easy". • distribution is given by µ and Σ. • linear combinations are normal (! VaR and ES calcs easy). • conditional distributions are normal. > • For (X1;X2) ∼ N2(µ; Σ), ρ(X1;X2) = 0 () X1 and X2 are independent: QRM 2010 75 Multivariate Normal Variance Mixtures Cons of Multivariate Normal Distribution • tails are thin, meaning that extreme values are scarce in the normal model. • joint extremes in the multivariate model are also too scarce. • the distribution has a strong form of symmetry, called elliptical symmetry. How to repair the drawbacks of the multivariate normal model? QRM 2010 76 Multivariate Normal Variance Mixtures The random vector X has a (multivariate) normal variance mixture distribution if d p X = µ + WAZ; (1) where • Z ∼ Nk(0;Ik); • W ≥ 0 is a scalar random variable which is independent of Z; and • A 2 Rd×k and µ 2 Rd are a matrix and a vector of constants, respectively. > Set Σ := AA . Observe: XjW = w ∼ Nd(µ; wΣ). QRM 2010 77 Multivariate Normal Variance Mixtures Assumption: rank(A)= d ≤ k, so Σ is a positive definite matrix. If E(W ) < 1 then easy calculations give E(X) = µ and cov(X) = E(W )Σ: We call µ the location vector or mean vector and we call Σ the dispersion matrix. The correlation matrices of X and AZ are identical: corr(X) = corr(AZ): Multivariate normal variance mixtures provide the most useful examples of elliptical distributions.
    [Show full text]
  • Mixture Modeling with Applications in Alzheimer's Disease
    University of Kentucky UKnowledge Theses and Dissertations--Epidemiology and Biostatistics College of Public Health 2017 MIXTURE MODELING WITH APPLICATIONS IN ALZHEIMER'S DISEASE Frank Appiah University of Kentucky, [email protected] Digital Object Identifier: https://doi.org/10.13023/ETD.2017.100 Right click to open a feedback form in a new tab to let us know how this document benefits ou.y Recommended Citation Appiah, Frank, "MIXTURE MODELING WITH APPLICATIONS IN ALZHEIMER'S DISEASE" (2017). Theses and Dissertations--Epidemiology and Biostatistics. 14. https://uknowledge.uky.edu/epb_etds/14 This Doctoral Dissertation is brought to you for free and open access by the College of Public Health at UKnowledge. It has been accepted for inclusion in Theses and Dissertations--Epidemiology and Biostatistics by an authorized administrator of UKnowledge. For more information, please contact [email protected]. STUDENT AGREEMENT: I represent that my thesis or dissertation and abstract are my original work. Proper attribution has been given to all outside sources. I understand that I am solely responsible for obtaining any needed copyright permissions. I have obtained needed written permission statement(s) from the owner(s) of each third-party copyrighted matter to be included in my work, allowing electronic distribution (if such use is not permitted by the fair use doctrine) which will be submitted to UKnowledge as Additional File. I hereby grant to The University of Kentucky and its agents the irrevocable, non-exclusive, and royalty-free license to archive and make accessible my work in whole or in part in all forms of media, now or hereafter known.
    [Show full text]
  • A Two-Component Normal Mixture Alternative to the Fay-Herriot Model
    A two-component normal mixture alternative to the Fay-Herriot model August 22, 2015 Adrijo Chakraborty 1, Gauri Sankar Datta 2 3 4 and Abhyuday Mandal 5 Abstract: This article considers a robust hierarchical Bayesian approach to deal with ran- dom effects of small area means when some of these effects assume extreme values, resulting in outliers. In presence of outliers, the standard Fay-Herriot model, used for modeling area- level data, under normality assumptions of the random effects may overestimate random effects variance, thus provides less than ideal shrinkage towards the synthetic regression pre- dictions and inhibits borrowing information. Even a small number of substantive outliers of random effects result in a large estimate of the random effects variance in the Fay-Herriot model, thereby achieving little shrinkage to the synthetic part of the model or little reduction in posterior variance associated with the regular Bayes estimator for any of the small areas. While a scale mixture of normal distributions with known mixing distribution for the random effects has been found to be effective in presence of outliers, the solution depends on the mixing distribution. As a possible alternative solution to the problem, a two-component nor- mal mixture model has been proposed based on noninformative priors on the model variance parameters, regression coefficients and the mixing probability. Data analysis and simulation studies based on real, simulated and synthetic data show advantage of the proposed method over the standard Bayesian Fay-Herriot solution derived under normality of random effects. 1NORC at the University of Chicago, Bethesda, MD 20814, [email protected] 2Department of Statistics, University of Georgia, Athens, GA 30602, USA.
    [Show full text]
  • Student's T Distribution Based Estimation of Distribution
    1 Student’s t Distribution based Estimation of Distribution Algorithms for Derivative-free Global Optimization ∗Bin Liu Member, IEEE, Shi Cheng Member, IEEE, Yuhui Shi Fellow, IEEE Abstract In this paper, we are concerned with a branch of evolutionary algorithms termed estimation of distribution (EDA), which has been successfully used to tackle derivative-free global optimization problems. For existent EDA algorithms, it is a common practice to use a Gaussian distribution or a mixture of Gaussian components to represent the statistical property of available promising solutions found so far. Observing that the Student’s t distribution has heavier and longer tails than the Gaussian, which may be beneficial for exploring the solution space, we propose a novel EDA algorithm termed ESTDA, in which the Student’s t distribution, rather than Gaussian, is employed. To address hard multimodal and deceptive problems, we extend ESTDA further by substituting a single Student’s t distribution with a mixture of Student’s t distributions. The resulting algorithm is named as estimation of mixture of Student’s t distribution algorithm (EMSTDA). Both ESTDA and EMSTDA are evaluated through extensive and in-depth numerical experiments using over a dozen of benchmark objective functions. Empirical results demonstrate that the proposed algorithms provide remarkably better performance than their Gaussian counterparts. Index Terms derivative-free global optimization, EDA, estimation of distribution, Expectation-Maximization, mixture model, Student’s t distribution I. INTRODUCTION The estimation of distribution algorithm (EDA) is an evolutionary computation (EC) paradigm for tackling derivative-free global optimization problems [1]. EDAs have gained great success and attracted increasing attentions during the last decade arXiv:1608.03757v2 [cs.NE] 25 Nov 2016 [2].
    [Show full text]
  • Metaplus: Robust Meta-Analysis and Meta-Regression
    Package ‘metaplus’ June 12, 2021 Type Package Title Robust Meta-Analysis and Meta-Regression Version 1.0-2 Date 2021-06-10 Maintainer Ken Beath <[email protected]> Contact Ken Beath <[email protected]> Author Ken Beath [aut, cre], Ben Bolker [aut], R Development Core Team [aut] Description Performs meta-analysis and meta-regression using standard and robust methods with con- fidence intervals based on the profile likelihood. Robust methods are based on alternative distri- butions for the random effect, either the t-distribution (Lee and Thomp- son, 2008 <doi:10.1002/sim.2897> or Baker and Jackson, 2008 <doi:10.1007/s10729-007-9041- 8>) or mixtures of normals (Beath, 2014 <doi:10.1002/jrsm.1114>). Depends R(>= 3.2.0) Suggests R.rsp VignetteBuilder R.rsp Imports bbmle, metafor, boot, methods, numDeriv, MASS, graphics, stats, fastGHQuad, lme4, Rfast, parallel, doParallel, foreach, doRNG LazyData yes License GPL (>= 2) NeedsCompilation no Repository CRAN Date/Publication 2021-06-12 04:20:02 UTC R topics documented: metaplus-package . .2 AIC .............................................3 1 2 metaplus-package BIC .............................................3 cdp..............................................4 exercise . .5 logLik . .6 mag .............................................6 marinho . .7 metaplus . .8 outlierProbs . 12 plot.metaplus . 13 plot.outlierProbs . 14 summary . 15 testOutliers . 16 Index 18 metaplus-package Fits random effects meta-analysis models including robust models Description Allows fitting of random effects meta-analysis producing confidence intervals based on the profile likelihood (Hardy and Thompson, 1996). Two methods of robust meta-analysis are included, based on either the t-distribution (Baker and Jackson (2008) and Lee and Thompson (2008)) or normal- mixture distribution (Beath, 2014).
    [Show full text]
  • Bayesian Random-Effects Meta-Analysis Using the Bayesmeta
    JSS Journal of Statistical Software MMMMMM YYYY, Volume VV, Issue II. doi: 10.18637/jss.v000.i00 Bayesian random-effects meta-analysis using the bayesmeta R package Christian R¨over University Medical Center G¨ottingen Abstract The random-effects or normal-normal hierarchical model is commonly utilized in a wide range of meta-analysis applications. A Bayesian approach to inference is very at- tractive in this context, especially when a meta-analysis is based only on few studies. The bayesmeta R package provides readily accessible tools to perform Bayesian meta-analyses and generate plots and summaries, without having to worry about computational details. It allows for flexible prior specification and instant access to the resulting posterior distri- butions, including prediction and shrinkage estimation, and facilitating for example quick sensitivity checks. The present paper introduces the underlying theory and showcases its usage. Keywords: evidence synthesis, NNHM, between-study heterogeneity. 1. Introduction 1.1. Meta-analysis arXiv:1711.08683v2 [stat.CO] 10 Oct 2018 Evidence commonly comes in separate bits, and not necessarily from a single experiment. In contemporary science, the careful conduct of systematic reviews of the available evidence from diverse data sources is an effective and ubiquitously practiced means of compiling relevant in- formation. In this context, meta-analyses allow for the formal, mathematical combination of information to merge data from individual investigations to a joint result. Along with qualita- tive, often informal assessment and evaluation of the present evidence, meta-analytic methods have become a powerful tool to guide objective decision-making (Chalmers et al. 2002; Liberati et al.
    [Show full text]
  • A Class of Scale Mixtured Normal Distribution Rezaul Karim Md* Department of Statistics, Jahangirnagar University, Bangladesh
    metrics & io B B i f o o s t l a a t i Rezaul Karim, J Biom Biostat 2016, 7:2 n s r t i u c o s J Journal of Biometrics & Biostatistics DOI: 10.4172/2155-6180.1000285 ISSN: 2155-6180 Research Article Article OpenOpen Access Access A Class of Scale Mixtured Normal Distribution Rezaul Karim Md* Department of Statistics, Jahangirnagar University, Bangladesh Abstract A class of scale mixtured normal distribution with lifetime probability distributions has been proposed in this article. Different moments, characteristic function and shape characteristics of these proposed probability distributions have also been provided. The scale mixture of normal distribution is extensively used in Biostatistical field. In this article, some new lifetime probability distributions have been proposed which is the scale mixture of normal distribution. Different properties such as moments, characteristic function, shape characteristics of these probability distributions have also been mentioned. Keyword: Scale mixture distribution; Inverted probability Main results distribution; Moments; Skewness; Kurtosis Here, the definition of scale mixture of normal distribution with Introduction well known lifetime distributions such as inverted chi-square, gamma, exponential, Rayleigh distributions are mentioned. The important A mixture distribution is a convex function of a set of probability characteristics of these distributions such as characteristic functions, distributions [1]. At first, Pearson considered a mixture distribution moments, and shape characteristics are also mentioned. The main of two normal distributions in 1894 and he studied the estimation results of the paper are presented in the form of definitions and of the parameters of his proposed distribution [2]. After around half theorems.
    [Show full text]
  • Derivation of Mixture Distributions and Weighted Likelihood Function As Minimizers of Kl-Divergence Subject to Constraints
    Ann. Inst. Statist. Math. Vol. 57, No. 4, 687 701 (2005) @2005 The Institute of Statistical Mathematics DERIVATION OF MIXTURE DISTRIBUTIONS AND WEIGHTED LIKELIHOOD FUNCTION AS MINIMIZERS OF KL-DIVERGENCE SUBJECT TO CONSTRAINTS XIAOGANG WANG 1 AND JAMES V. ZIDEK 2 1Department of Mathematics and Statistics, York University, 4700 Keele Street, ON, Canada M3J 1P3 2Department of Statistics, University of British Columbia, 333-6356 Agriculture Road, BC, Canada V6T 1Z2 (Received May 26, 2003; revised July 7, 2004) Abstract. In this article, mixture distributions and weighted likelihoods are de- rived within an information-theoretic framework and shown to be closely related. This surprising relationship obtains in spite of the arithmetic form of the former and the geometric form of the latter. Mixture distributions are shown to be optima that minimize the entropy loss under certain constraints. The same framework implies the weighted likelihood when the distributions in the mixture are unknown and informa- tion from independent samples generated by them have to be used instead. Thus the likelihood weights trade bias for precision and yield inferential procedures such as estimates that can be more reliable than their classical counterparts. Key words and phrases: Euler-Lagrange equations, relative entropy, mixture distri- butions, weighted likelihood. 1. Introduction To introduce the results we present in this paper, suppose a statistician, S, is required to predict y, the realized value of a random variable Y with unknown probability density function (PDF), say f. Instead of a point prediction, S is allowed to state the prediction as a predictive PDF, say g. Finally, S receives a reward, log g(y).
    [Show full text]
  • Multivariate Distributions
    IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing1 in particular on multivariate normal, normal-mixture, spherical and elliptical distributions. In addition to studying their properties, we will also discuss techniques for simulating and, very briefly, estimating these distributions. Familiarity with these important classes of multivariate distributions is important for many aspects of risk management. We will defer the study of copulas until later in the course. 1 Preliminary Definitions Let X = (X1;:::Xn) be an n-dimensional vector of random variables. We have the following definitions and statements. > n Definition 1 (Joint CDF) For all x = (x1; : : : ; xn) 2 R , the joint cumulative distribution function (CDF) of X satisfies FX(x) = FX(x1; : : : ; xn) = P (X1 ≤ x1;:::;Xn ≤ xn): Definition 2 (Marginal CDF) For a fixed i, the marginal CDF of Xi satisfies FXi (xi) = FX(1;:::; 1; xi; 1;::: 1): It is straightforward to generalize the previous definition to joint marginal distributions. For example, the joint marginal distribution of Xi and Xj satisfies Fij(xi; xj) = FX(1;:::; 1; xi; 1;:::; 1; xj; 1;::: 1). If the joint CDF is absolutely continuous, then it has an associated probability density function (PDF) so that Z x1 Z xn FX(x1; : : : ; xn) = ··· f(u1; : : : ; un) du1 : : : dun: −∞ −∞ Similar statements also apply to the marginal CDF's. A collection of random variables is independent if the joint CDF (or PDF if it exists) can be factored into the product of the marginal CDFs (or PDFs). If > > X1 = (X1;:::;Xk) and X2 = (Xk+1;:::;Xn) is a partition of X then the conditional CDF satisfies FX2jX1 (x2jx1) = P (X2 ≤ x2jX1 = x1): If X has a PDF, f(·), then it satisfies Z xk+1 Z xn f(x1; : : : ; xk; uk+1; : : : ; un) FX2jX1 (x2jx1) = ··· duk+1 : : : dun −∞ −∞ fX1 (x1) where fX1 (·) is the joint marginal PDF of X1.
    [Show full text]