Statistical Pattern Recognition: a Review

Total Page:16

File Type:pdf, Size:1020Kb

Statistical Pattern Recognition: a Review 4 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 22, NO. 1, JANUARY 2000 Statistical Pattern Recognition: A Review Anil K. Jain, Fellow, IEEE, Robert P.W. Duin, and Jianchang Mao, Senior Member, IEEE AbstractÐThe primary goal of pattern recognition is supervised or unsupervised classification. Among the various frameworks in which pattern recognition has been traditionally formulated, the statistical approach has been most intensively studied and used in practice. More recently, neural network techniques and methods imported from statistical learning theory have been receiving increasing attention. The design of a recognition system requires careful attention to the following issues: definition of pattern classes, sensing environment, pattern representation, feature extraction and selection, cluster analysis, classifier design and learning, selection of training and test samples, and performance evaluation. In spite of almost 50 years of research and development in this field, the general problem of recognizing complex patterns with arbitrary orientation, location, and scale remains unsolved. New and emerging applications, such as data mining, web searching, retrieval of multimedia data, face recognition, and cursive handwriting recognition, require robust and efficient pattern recognition techniques. The objective of this review paper is to summarize and compare some of the well-known methods used in various stages of a pattern recognition system and identify research topics and applications which are at the forefront of this exciting and challenging field. Index TermsÐStatistical pattern recognition, classification, clustering, feature extraction, feature selection, error estimation, classifier combination, neural networks. æ 1INTRODUCTION Y the time they are five years old, most children can depend in some way on pattern recognition... We need to Brecognize digits and letters. Small characters, large pay much more explicit attention to teaching pattern characters, handwritten, machine printed, or rotatedÐall recognition.º Our goal here is to introduce pattern recogni- are easily recognized by the young. The characters may be tion as the best possible way of utilizing available sensors, written on a cluttered background, on crumpled paper or processors, and domain knowledge to make decisions may even be partially occluded. We take this ability for automatically. granted until we face the task of teaching a machine how to do the same. Pattern recognition is the study of how 1.1 What is Pattern Recognition? machines can observe the environment, learn to distinguish Automatic (machine) recognition, description, classifica- patterns of interest from their background, and make sound tion, and grouping of patterns are important problems in a and reasonable decisions about the categories of the variety of engineering and scientific disciplines such as patterns. In spite of almost 50 years of research, design of biology, psychology, medicine, marketing, computer vision, a general purpose machine pattern recognizer remains an artificial intelligence, and remote sensing. But what is a elusive goal. pattern? Watanabe [163] defines a pattern ªas opposite of a The best pattern recognizers in most instances are chaos; it is an entity, vaguely defined, that could be given a humans, yet we do not understand how humans recognize name.º For example, a pattern could be a fingerprint image, patterns. Ross [140] emphasizes the work of Nobel Laureate a handwritten cursive word, a human face, or a speech Herbert Simon whose central finding was that pattern signal. Given a pattern, its recognition/classification may recognition is critical in most human decision making tasks: consist of one of the following two tasks [163]: 1) supervised ªThe more relevant patterns at your disposal, the better classification (e.g., discriminant analysis) in which the input your decisions will be. This is hopeful news to proponents pattern is identified as a member of a predefined class, 2) unsupervised classification (e.g., clustering) in which the of artificial intelligence, since computers can surely be pattern is assigned to a hitherto unknown class. Note that taught to recognize patterns. Indeed, successful computer the recognition problem here is being posed as a classifica- programs that help banks score credit applicants, help tion or categorization task, where the classes are either doctors diagnose disease and help pilots land airplanes defined by the system designer (in supervised classifica- tion) or are learned based on the similarity of patterns (in . A.K. Jain is with the Department of Computer Science and Engineering, unsupervised classification). Michigan State University, East Lansing, MI 48824. Interest in the area of pattern recognition has been E-mail: [email protected]. renewed recently due to emerging applications which are . R.P.W. Duin is with the Department of Applied Physics, Delft University of Technology, 2600 GA Delft, the Netherlands. not only challenging but also computationally more E-mail: [email protected]. demanding (see Table 1). These applications include data . J. Mao is with the IBM Almaden Research Center, 650 Harry Road, San mining (identifying a ªpattern,º e.g., correlation, or an Jose, CA 95120. E-mail: [email protected]. outlier in millions of multidimensional patterns), document Manuscript received 23 July 1999; accepted 12 Oct. 1999. classification (efficiently searching text documents), finan- Recommended for acceptance by K. Bowyer. For information on obtaining reprints of this article, please send e-mail to: cial forecasting, organization and retrieval of multimedia [email protected], and reference IEEECS Log Number 110296. databases, and biometrics (personal identification based on 0162-8828/00/$10.00 ß 2000 IEEE JAIN ET AL.: STATISTICAL PATTERN RECOGNITION: A REVIEW 5 TABLE 1 Examples of Pattern Recognition Applications various physical attributes such as face and fingerprints). well-defined and sufficiently constrained recognition pro- Picard [125] has identified a novel application of pattern blem (small intraclass variations and large interclass recognition, called affective computing which will give a variations) will lead to a compact pattern representation computer the ability to recognize and express emotions, to and a simple decision making strategy. Learning from a set respond intelligently to human emotion, and to employ of examples (training set) is an important and desired mechanisms of emotion that contribute to rational decision attribute of most pattern recognition systems. The four best making. A common characteristic of a number of these known approaches for pattern recognition are: 1) template applications is that the available features (typically, in the matching, 2) statistical classification, 3) syntactic or struc- thousands) are not usually suggested by domain experts, tural matching, and 4) neural networks. These models are but must be extracted and optimized by data-driven not necessarily independent and sometimes the same procedures. pattern recognition method exists with different interpreta- The rapidly growing and available computing power, tions. Attempts have been made to design hybrid systems while enabling faster processing of huge data sets, has also involving multiple models [57]. A brief description and facilitated the use of elaborate and diverse methods for data comparison of these approaches is given below and analysis and classification. At the same time, demands on summarized in Table 2. automatic pattern recognition systems are rising enor- 1.2 Template Matching mously due to the availability of large databases and One of the simplest and earliest approaches to pattern stringent performance requirements (speed, accuracy, and recognition is based on template matching. Matching is a cost). In many of the emerging applications, it is clear that generic operation in pattern recognition which is used to no single approach for classification is ªoptimalº and that determine the similarity between two entities (points, multiple methods and approaches have to be used. curves, or shapes) of the same type. In template matching, Consequently, combining several sensing modalities and a template (typically, a 2D shape) or a prototype of the classifiers is now a commonly used practice in pattern pattern to be recognized is available. The pattern to be recognition. recognized is matched against the stored template while The design of a pattern recognition system essentially taking into account all allowable pose (translation and involves the following three aspects: 1) data acquisition and rotation) and scale changes. The similarity measure, often a preprocessing, 2) data representation, and 3) decision correlation, may be optimized based on the available making. The problem domain dictates the choice of training set. Often, the template itself is learned from the sensor(s), preprocessing technique, representation scheme, training set. Template matching is computationally de- and the decision making model. It is generally agreed that a manding, but the availability of faster processors has now 6 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 22, NO. 1, JANUARY 2000 TABLE 2 Pattern Recognition Models made this approach more feasible. The rigid template 1.4 Syntactic Approach matching mentioned above, while effective in some In many recognition problems involving complex patterns, application domains, has a number of disadvantages. For
Recommended publications
  • Crash Prediction and Collision Avoidance Using Hidden
    CRASH PREDICTION AND COLLISION AVOIDANCE USING HIDDEN MARKOV MODEL A Thesis Submitted to the Faculty of Purdue University by Avinash Prabu In Partial Fulfillment of the Requirements for the Degree of Master of Science in Electrical and Computer Engineering August 2019 Purdue University Indianapolis, Indiana ii THE PURDUE UNIVERSITY GRADUATE SCHOOL STATEMENT OF THESIS APPROVAL Dr. Lingxi Li, Chair Department of Electrical and Computer Engineering Dr. Yaobin Chen Department of Electrical and Computer Engineering Dr. Brian King Department of Electrical and Computer Engineering Approved by: Dr. Brian King Head of Graduate Program iii Dedicated to my family, friends and my Professors, for their support and encouragement. iv ACKNOWLEDGMENTS Firstly, I would like to thank Dr. Lingxi Li for his continuous support and en- couragement. He has been a wonderful advisor and an exceptional mentor, without whom, my academic journey will not be where it is now. I would also like to to thank Dr. Yaobin Chen, whose course on Control theory has had a great impact in shaping this thesis. I would like to extend my appreciation to Dr. Brian King, who has been a great resource of guidance throughout the course of my Master's degree. A special thanks to Dr. Renran Tian, who is my principle investigator at TASI, IUPUI, for his guidance and encouragement. I also wish to thank Sherrie Tucker, who has always been available to answer my questions, despite her busy schedule. I would like to acknowledge my parents, Mr. & Mrs. Prabhu and my sister Chathurma for their continuous belief and endless love towards me.
    [Show full text]
  • Probabilistic Sequence Models with Speech and Language Applications
    Thesis for the degree of Doctor of Philosophy Probabilistic Sequence Models with Speech and Language Applications Gustav Eje Henter Communication Theory School of Electrical Engineering KTH – Royal Institute of Technology Stockholm 2013 Henter, Gustav Eje Probabilistic Sequence Models with Speech and Language Applications Copyright c 2013 Gustav Eje Henter except where otherwise stated. All rights reserved. ISBN 978-91-7501-932-1 TRITA-EE 2013:042 ISSN 1653-5146 Communication Theory School of Electrical Engineering KTH – Royal Institute of Technology SE-100 44 Stockholm, Sweden Abstract Series data, sequences of measured values, are ubiquitous. Whenever ob- servations are made along a path in space or time, a data sequence results. To comprehend nature and shape it to our will, or to make informed de- cisions based on what we know, we need methods to make sense of such data. Of particular interest are probabilistic descriptions, which enable us to represent uncertainty and random variation inherent to the world around us. This thesis presents and expands upon some tools for creating prob- abilistic models of sequences, with an eye towards applications involving speech and language. Modelling speech and language is not only of use for creating listening, reading, talking, and writing machines—for instance al- lowing human-friendly interfaces to future computational intelligences and smart devices of today—but probabilistic models may also ultimately tell us something about ourselves and the world we occupy. The central theme of the thesis is the creation of new or improved models more appropriate for our intended applications, by weakening limiting and questionable assumptions made by standard modelling techniques.
    [Show full text]
  • Unsupervised Language Acquisition: Theory and Practice
    Unsupervised Language Acquisition: Theory and Practice Alexander Simon Clark Submitted for the degree of D.Phil. University of Sussex September, 2001 Declaration I hereby declare that this thesis has not been submitted, either in the same or different form, to this or any other university for a degree. Signature: Acknowledgements First, I would like to thank Bill Keller, for his supervision over the past three years. I would like to thank my colleagues at ISSCO for making me welcome and especially Susan Armstrong for giving me the job in the first place, and various people for helpful comments and discussions including Chris Manning, Dan Klein, Eric Gaussier, Nicola Cancedda, Franck Thollard, Andrei Popescu-Beilis, Menno van Zaanen and numerous other people including Sonia Halimi for check- ing my Arabic. I would also like to thank Gerald Gazdar and Geoffrey Sampson for their helpful comments as part of my thesis committee. I would also like to thank all of the people that have worked on the various software packages and operating systems that I have used, such as LaTeX, Gnu/Linux, and the gcc compiler, as well as the people who have worked on the preparation of the various corpora I have used: in particular I am grateful to John McCarthy and Ramin Nakisa for allowing me to use their painstakingly collected data sets for Arabic. Finally I would like to thank my wife, Dr Clare Hornsby, for her unfailing support and encouragement (in spite of her complete lack of understanding of the subject) and my children, Lily and Francis, for putting up with my absences both physical and mental, and showing me how language acquisition is really done – over-regularisation of past tense verbs included.
    [Show full text]
  • Simplicity: a Unifying Principle in Cognitive Science?
    Simplicity: A unifying principle in cognitive science? Nick Chater Institute for Applied Cognitive Science Department of Psychology University of Warwick Coventry, CV4 7AL UK Paul Vitányi, Centrum voor Wiskunde en Informatica Kruislaan 413 1098 SJ Amsterdam The Netherlands 1 Summary Much of perception, learning and high-level cognition involves finding patterns in data. But there are always infinitely many patterns compatible with any finite amount of data. How does the cognitive system choose ‘sensible’ patterns? A long tradition in epistemology, philosophy of science, and mathematical and computational theories of learning argues that patterns ‘should’ be chosen according to how simply they explain the data. This article reviews research exploring the idea that simplicity does, indeed, drive a wide range of cognitive processes. We outline mathematical theory, computational results, and empirical data underpinning this viewpoint. Key words: simplicity, Kolmogorov complexity, codes, learning, induction, Bayesian inference 30-word summary:This article outlines the proposal that many aspects of cognition, from perception, to language acquisition, to high-level cognition involve finding patterns that provide the simplest explanation of available data. 2 The cognitive system finds patterns in the data that it receives. Perception involves finding patterns in the external world, from sensory input. Language acquisition involves finding patterns in linguistic input, to determine the structure of the language. High-level cognition involves finding patterns in information, to form categories, and to infer causal relations. Simplicity and the problem of induction A fundamental puzzle is what we term the problem of induction: infinitely many patterns are compatible with any finite set of data (see Box 1).
    [Show full text]
  • CSCI 5521: Pattern Recognition Prof
    CSCI 5521: Pattern Recognition Prof. Paul Schrater Business • Course web page: http://gandalf.psych.umn.edu/~schrater/schrater_lab/courses /PattRecog07/PattRecog.html Prof. Paul Schrater Pattern Recognition CSCI 5521 2 Syllabus • Reading materials: – Pattern Classification, 2nd Ed. Duda, Hart, Stork – Statistical Pattern Recognition, 2nd Ed. Andrew Webb – Elements of Statistical Learning: Data Mining, Inference, Prediction, Trevor Hastie, Robert Tibshirani, and Jerome Friedman. Springer, 2001. – Gaussian Processes for Machine Learning, Carl Rasmussen and Christopher K.I. Williams, MIT Press, 2006. – Papers posted on the web site. – Downloads will be password protected. • Grading • 60% on the homework assignments • 40% on the final project. Prof. Paul Schrater Pattern Recognition CSCI 5521 3 Syllabus cont’d • Final Project 12-15 page paper involving: 1) Simulation or experiments. For example, implement a pattern recognition system for a particular application, e.g. digit classification, document clustering, etc. 2) Literature survey (with critical evaluation) on a given topic. 3) Theoretical work (detailed derivations, extensions of existing work, etc) Important dates: • Sept. 28: Topic selection. One or two pages explaining the project with a list of references. • Nov. 7: Partial report (3 to 5 pages). • Dec. 20: Final report (12 to 15 pages). • Students may work in groups of 2-4. Prof. Paul Schrater Pattern Recognition CSCI 5521 4 Policies/Procedures • DO NOT CHEAT. • Do NOT work in groups for homework (consulting with fellow students permitted, but group submissions are not). • Electronically submit homework. • Homework must be submitted by Midnight on the day it is due. Prof. Paul Schrater Pattern Recognition CSCI 5521 5 Introduction to Pattern Recognition • Syllabus • What are Patterns? • Pattern Recognition • An Example • Pattern Recognition Systems • The Design Cycle • Learning and Adaptation • Conclusion Examples of Patterns Prof.
    [Show full text]
  • Epurdy-Phd-Dissertation.Pdf
    THE UNIVERSITY OF CHICAGO GRAMMATICAL METHODS IN COMPUTER VISION A DISSERTATION SUBMITTED TO THE FACULTY OF THE DIVISION OF THE PHYSICAL SCIENCES IN CANDIDACY FOR THE DEGREE OF DOCTOR OF PHILOSOPHY DEPARTMENT OF COMPUTER SCIENCE BY ERIC PURDY CHICAGO, ILLINOIS MARCH 2013 Copyright c 2013 by Eric Purdy All Rights Reserved To my beloved wife TABLE OF CONTENTS LIST OF FIGURES . vii LIST OF TABLES . xii ABSTRACT . xiii 1 INTRODUCTION . 1 1.1 Examples of Grammatical Approaches . .2 1.1.1 Curve Grammars . .2 1.1.2 Visual Chalkboard . .4 1.1.3 Visual Search Engine . .4 1.1.4 Other Visual Grammars . .4 1.2 Grammatical Vision is Important . .6 1.2.1 Soft Decisions . .6 1.2.2 Modeling Clutter with Object Sub-parts . .7 1.2.3 Whole Scene Parsing . .7 1.3 Hierarchical Decomposition and Rich Description . .8 1.3.1 Training on rich annotations . .9 1.3.2 Some Problems with Rich Description, and Solutions . 10 1.3.3 Rich Description and XML . 12 1.4 Grammars and Statistical Models . 13 1.4.1 Statistical Modules . 13 1.4.2 Independence and the Poverty of Stimulus . 13 1.5 Grammatical Vision is Difficult . 15 1.6 Grammar induction is difficult . 15 1.7 Visual Grammars are hard for additional reasons . 16 1.7.1 Comparisons are Always Inexact . 16 1.7.2 Dealing with Scale . 17 1.7.3 Plane grammars do not naturally admit efficient parsing . 18 1.8 Contributions . 18 2 GRAMMATICAL MODELS OF SHAPE . 20 2.1 Introduction . 20 2.2 Motivating Example: L-Systems .
    [Show full text]
  • Cveigh, E., Kass, D., Miller, M.I., Winslow, R.L
    Curriculum Vitae Michael I. Miller https://en.wikipedia.org/wiki/Michael_I._Miller Bessie Darling Massey Professor and Director, Johns Hopkins Department of Biomedical Engineering Director, Center for Imaging Science Co-Director, Kavli Neuroscience Discovery Institute Gilman Scholar Professor of Electrical and Computer Engineering (410) 516-3826 (office) [email protected] Education Ph.D. Biomedical Engineering, Johns Hopkins University, 1984 M.S. Electrical Engineering, Johns Hopkins University, 1979 B.S. Electrical Engineering, State University of New York at Stony Brook, 1976 Experience • Bessie Darling Massey Professor and Director, Department of Biomedical Engineering. 2017 - Present • Co-Director, Kavli Neuroscience Discovery Institute. 2015 – Present • Director, Center for Imaging Science. 2000 - Present • Gilman Scholar, Johns Hopkins University. 2011 – Present • Herschel and Ruth Seder Chair in Biomedical Engineering, Johns Hopkins University. 2004 - 2017 • Professor, Johns Hopkins University. 1998 - Present • Visiting Professor,École Normale Supérieure de Cachan, France. 2000, May 2002, May 2007, May 2009, May 2011, June 2014, June 2015 • Visiting Professor, Institute Henri Poincaré, France. November, 1998 • Visiting Professor, Brown University. 1990,1994,1995-2001 • Newton R. and Sarah L. Wilson Chair in Biomedical Engineering, Washington University. 1995-1998 • Professor, Washington University St. Louis. 1992 - 1998 • Visiting Professor, Université René Descartes. November. 1997 • Associate Professor, Washington University St.
    [Show full text]
  • Generating Open World Descriptions of Video Using Commonsense
    QUARTERLY OF APPLIED MATHEMATICS VOLUME LXXVII, NUMBER 2 JUNE 2019, PAGES 323–356 https://doi.org/10.1090/qam/1530 Article electronically published on January 11, 2019 GENERATING OPEN WORLD DESCRIPTIONS OF VIDEO USING COMMON SENSE KNOWLEDGE IN A PATTERN THEORY FRAMEWORK By SATHYANARAYANAN N. AAKUR (Department of Computer Science and Engineering, University of South Florida, Tampa, Florida 33620 ), FILLIPE DM DE SOUZA (Department of Computer Science and Engineering, University of South Florida, Tampa, Florida 33620), and SUDEEP SARKAR (Department of Computer Science and Engineering, University of South Florida, Tampa, Florida 33620 ) This paper is dedicated to Professor Ulf Grenander Abstract. The task of interpretation of activities as captured in video extends beyond just the recognition of observed actions and objects. It involves open world reasoning and constructing deep semantic connections that go beyond what is directly observed in the video and annotated in the training data. Prior knowledge plays a big role. Grenander’s canonical pattern theory representation offers an elegant mechanism to capture these semantic connections between what is observed directly in the image and past knowl- edge in large-scale common sense knowledge bases, such as ConceptNet. We represent interpretations using a connected structure of basic detected (grounded) concepts, such as objects and actions, that are bound by semantics with other background concepts not directly observed, i.e., contextualization cues. Concepts are basic generators and the bonds are defined by the semantic relationships between concepts. Local and global regularity constraints govern these bonds and the overall connection structure. We use an inference engine based on energy minimization using an efficient Markov Chain Monte Carlo that uses the ConceptNet in its move proposals to find these structures that de- scribe the image content.
    [Show full text]
  • NEW BOOKS Pattern Theory: the Stochastic Analysis of Real-World
    QUARTERLY OF APPLIED MATHEMATICS Volume 69, Number 2, June 2011, Pages 403–404 S 0033-569X(2011)01270-6 Article electronically published on March 23, 2011 NEW BOOKS Pattern Theory: The stochastic analysis of real-world signals. By David Mumford and Ag- nes Desolneux, A K Peters, Ltd., 407 pp., $79.00 This is a remarkable work. It is full of sophisticated mathematical models dealing with all sorts of phenomena from the physical as well as from the man-made world. Or, to use the author’s parlance, “analysis of real-world signals”. Much work in statistics, pattern recognition, etc. is based on highly simplified models and reflect more the researcher’s statistical expertise than an attempt at real understanding of the phenomena studies. In contrast, this book goes deeper and presents serious attempts to really dig deeply into the subject matter of the signals being analyzed. At the same time it does not avoid the mathematical difficulties in the construction of reality-based models. Indeed, each chapter contains a section called Basics, in which the reader is being prepared to use the mathematical tools suitable for a particular application. These tools are often of an advanced character. Even elementary ideas are presented from a higher and unifying perspective. Or to quote Felix Klein: vom h¨oheren Standpunkt aus. While expertly guiding the reader to pattern theoretic thinking, this book also entertains by presenting additional material in the form of some delightful mathematics in the Basics sections. One of the first types of signals to be studied in an early chapter is English text (written) beginning with Claude Shannon’s thought experiment with simulated text.
    [Show full text]
  • Pattern Theory: the Mathematics of Perception
    ICM 2002 · Vol. III · 1–3 Pattern Theory: the Mathematics of Perception David Mumford∗ 1. Introduction How can we understand intelligent behavior? How can we design intelligent computers? These are questions that have been discussed by scientists and the public at large for over 50 years. As mathematicians, however, the question we want to ask is “is there a mathematical theory underlying intelligence?” I believe the first mathematical attack on these issues was Control Theory, led by Wiener and Pontryagin. They were studying how to design a controller which drives a motor affecting the world and also sits in a feedback loop receiving measurements from the world about the effect of the motor action. The goal was to control the motor so that the world, as measured, did something specific, i.e. move the tiller so that the boat stays on course. The main complication is that nothing is precisely predictable: the motor control is not exact, the world does unexpected things because of its complexities and the measurements you take of it are imprecise. All this led, in the simplest case, to a beautiful analysis known as the Wiener-Kalman-Bucy filter (to be described below). But Control Theory is basically a theory of the output side of intelligence with the measurements modeled in the simplest possible way: e.g. linear functions of the state of the world system being controlled plus additive noise. The real input side of intelligence is perception in a much broader sense, the analysis of all the noisy incomplete signals which you can pick up from the world through natural or artificial senses.
    [Show full text]
  • Discriminative, Descriptive and Generative Models
    QUARTERLY OF APPLIED MATHEMATICS VOLUME , NUMBER 0 XXXX XXXX, PAGES 000–000 A TALE OF THREE PROBABILISTIC FAMILIES: DISCRIMINATIVE, DESCRIPTIVE AND GENERATIVE MODELS BY YING NIAN WU (Department of Statistics, University of California, Los Angeles), RUIQI GAO (Department of Statistics, University of California, Los Angeles), TIAN HAN (Department of Statistics, University of California, Los Angeles), AND SONG-CHUN ZHU (Department of Statistics, University of California, Los Angeles) Abstract. The pattern theory of Grenander is a mathematical framework where patterns are represented by probability models on random variables of algebraic structures. In this paper, we review three families of probability models, namely, the discriminative models, the descriptive models, and the generative models. A discriminative model is in the form of a classifier. It specifies the con- ditional probability of the class label given the input signal. A descriptive model specifies the probability distribution of the signal, based on an energy function defined on the signal. A gen- erative model assumes that the signal is generated by some latent variables via a transformation. We shall review these models within a common framework and explore their connections. We shall also review the recent developments that take advantage of the high approximation capaci- ties of deep neural networks. 1. Introduction. Initially developed by Grenander in the 1970s, the pattern theory [30, 31] is a unified mathematical framework for representing, learning and recognizing patterns that arise in science and engineering. The objects in pattern theory are usually of high complexity or dimensionality, defined in terms of the constituent elements and the bonds between them. The patterns of these objects are characterized by both the algebraic structures governed by local and global rules, as well as the probability distributions of the associated random variables.
    [Show full text]
  • Markov Chain Monte Carlo Methods for Statistical Inference Julian Besag1 Department of Statistics University of Washington, Seattle, USA
    Markov Chain Monte Carlo Methods for Statistical Inference Julian Besag1 Department of Statistics University of Washington, Seattle, USA Spring 2004 SUMMARY These notes provide an introduction to Markov chain Monte Carlo methods and their applications to both Bayesian and frequentist statistical inference. Such methods have revolutionized what can be achieved computationally, es- pecially in the Bayesian paradigm. The account begins by discussing ordi- nary Monte Carlo methods: these have the same goals as the Markov chain versions but can only rarely be implemented. Subsequent sections describe basic Markov chain Monte Carlo, based on the Hastings algorithm and in- cluding both the Metropolis method and the Gibbs sampler as special cases, and go on to discuss some more specialized developments, including adaptive slice sampling, exact goodness{of{¯t tests, maximum likelihood estimation, the Langevin{Hastings algorithm, auxiliary variables techniques, perfect sam- pling via coupling from the past, reversible jumps methods for target spaces of varying dimensions, and simulated annealing. Specimen applications are described throughout the notes. Keywords: Adaptive slice sampling; Autologistic distribution; Auxiliary variables; Bayesian compu- tation; Competing risks; Contingency tables; Darwin's ¯nches; Exact p{values; Gibbs sampler; Hastings algorithm; Hidden Markov models; Importance sampling; Ising model; Langevin di®usion; Markov chain Monte Carlo; Markov random ¯elds; Maximum likelihood estimation; Metropolis method; Mixture mod- els;
    [Show full text]