Stochastic And-Or Grammars: a Unified Framework and Logic Perspective

Total Page:16

File Type:pdf, Size:1020Kb

Stochastic And-Or Grammars: a Unified Framework and Logic Perspective Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) Stochastic and-or Grammars: A Unified Framework and Logic Perspective⇤ Kewei Tu School of Information Science and Technology ShanghaiTech University, Shanghai, China [email protected] Abstract paper we will focus on the context-free subclass of stochas- tic AOGs, which serves as the skeleton in building more ad- Stochastic And-Or grammars (AOG) extend tradi- vanced stochastic AOGs. tional stochastic grammars of language to model other types of data such as images and events. In Several variants of stochastic AOGs and their inference al- this paper we propose a representation framework gorithms have been proposed in the literature to model differ- of stochastic AOGs that is agnostic to the type of ent types of data and solve different problems, such as image the data being modeled and thus unifies various scene parsing [Zhao and Zhu, 2011] and video event parsing domain-specific AOGs. Many existing grammar [Pei et al., 2011]. Our first contribution in this paper is that formalisms and probabilistic models in natural lan- we provide a unified representation framework of stochastic guage processing, computer vision, and machine AOGs that is agnostic to the type of the data being modeled; learning can be seen as special cases of this frame- in addition, based on this framework we propose a domain- work. We also propose a domain-independent in- independent inference algorithm that is tractable under a rea- ference algorithm of stochastic context-free AOGs sonable assumption. The benefits of a unified framework of and show its tractability under a reasonable as- stochastic AOGs include the following. First, such a frame- sumption. Furthermore, we provide two interpre- work can help us generalize and improve existing ad hoc ap- tations of stochastic context-free AOGs as a sub- proaches for modeling, inference and learning with stochastic set of probabilistic logic, which connects stochas- AOGs. Second, it also facilitates applications of stochastic tic AOGs to the field of statistical relational learn- AOGs to novel data types and problems and enables the re- ing and clarifies their relation with a few existing search of general-purpose inference and learning algorithms statistical relational models. of stochastic AOGs. Further, a formal definition of stochas- tic AOGs as abstract probabilistic models makes it easier to theoretically examine their relation with other models such 1 Introduction as constraint-based grammar formalism [Shieber, 1992] and Formal grammars are a popular class of knowledge represen- sum-product networks [Poon and Domingos, 2011]. In fact, tation that is traditionally confined to the modeling of natu- we will show that many of these related models can be seen ral and computer languages. However, several extensions of as special cases of stochastic AOGs. grammars have been proposed over time to model other types of data such as images [Fu, 1982; Zhu and Mumford, 2006; Stochastic AOGs model compositional structures based on Jin and Geman, 2006] and events [Ivanov and Bobick, 2000; the relations between sub-patterns. Such probabilistic model- Ryoo and Aggarwal, 2006; Pei et al., 2011]. One promi- ing of relational structures is traditionally studied in the field [ ] nent type of extension is stochastic And-Or grammars (AOG) of statistical relational learning Getoor and Taskar, 2007 . [Zhu and Mumford, 2006]. A stochastic AOG simultane- Our second contribution is that we provide probabilistic logic ously models compositions (i.e., a large pattern is the com- interpretations of the unified representation framework of position of several small patterns arranged according to a cer- stochastic AOGs and thus show that stochastic AOGs can tain configuration) and reconfigurations (i.e., a pattern may be seen as a novel type of statistical relational models. The have several alternative configurations), and in this way it can logic interpretations help clarify the relation between stochas- compactly represent a probabilistic distribution over a large tic AOGs and a few existing statistical relational models and number of patterns. Stochastic AOGs can be used to parse probabilistic logics that share certain features with stochas- [ data samples into their compositional structures, which help tic AOGs (e.g., tractable Markov logic Domingos and Webb, ] [ ] solve multiple tasks (such as classification, annotation, and 2012 and stochastic logic programs Muggleton, 1996 ). It segmentation of the data samples) in a unified manner. In this may also facilitate the incorporation of ideas from statistical relational learning into the study of stochastic AOGs and at ⇤This work was supported by the National Natural Science Foun- the same time contribute to the research of novel (tractable) dation of China (61503248). statistical relational models. 2654 2 Stochastic And-Or Grammars of non-overlapping sub-patterns. The And-rule speci- An AOG is an extension of a constituency grammar used in fies a production r : A x1,x2,...,xn for some n 2, where A is an And-node!{ and x ,x ,...,x} are natural language parsing [Manning and Schutze,¨ 1999]. Sim- ≥ 1 2 n ilar to a constituency grammar, an AOG defines a set of valid a set of terminal or nonterminal nodes representing the hierarchical compositions of atomic entities. However, an sub-patterns. A relation between the parameters of the child nodes, t(✓ ,✓ ,...,✓ ), specifies valid config- AOG differs from a constituency grammar in that it allows x1 x2 xn atomic entities other than words and compositional relations urations of the sub-patterns. This so-called parameter other than string concatenation. A stochastic AOG models relation is typically factorized to the conjunction of a the uncertainty in the composition by defining a probabilistic set of binary relations. A parameter function f is also distribution over the set of valid compositions. associated with the And-rule specifying how the param- Stochastic AOGs were first proposed to model images [Zhu eter of the And-node A is related to the parameters of the child nodes: ✓ = f(✓ ,✓ ,...,✓ ). We require that and Mumford, 2006; Zhao and Zhu, 2011; Wang et al., 2013; A x1 x2 xn Rothrock et al., 2013], in particular the spatial composition both the parameter relation and the parameter function of objects and scenes from atomic visual words (e.g., Garbor take time polynomial in n and m✓ to compute. There is bases). They were later extended to model events, in par- exactly one And-rule that is headed by each And-node. ticular the temporal and causal composition of events from An Or-rule, parameterized by an ordered pair r, p , rep- atomic actions [Pei et al., 2011] and fluents [Fire and Zhu, • resents an alternative configuration of a pattern.h i The 2013]. More recently, these two types of AOGs were used Or-rule specifies a production r : O x, where O jointly to model objects, scenes and events from the simulta- is an Or-node and x is either a terminal! or a nontermi- neous input of video and text [Tu et al., 2014]. nal node representing a possible configuration. A condi- In each of the previous work using stochastic AOGs, a tional probability p is associated with the Or-rule spec- different type of data is modeled with domain-specific and ifying how likely the configuration represented by x is problem-specific definitions of atomic entities and composi- selected given the Or-node O. The only constraint in the tions. Tu et al. [Tu et al., 2013] provided a first attempt to- Or-rule is that the parameters of O and x must be the wards a more unified definition of stochastic AOGs that is same: ✓O = ✓x. There typically exist multiple Or-rules agnostic to the type of the data being modeled. We refine and headed by the same Or-node, and together they can be extend their work by introducing parameterized patterns and written as O x1 x2 ... xn. relations in the unified definition, which allows us to reduce ! | | | a wide range of related models to AOGs (as will be discussed Note that unlike in some previous work, in the definition in section 2.1). Based on the unified framework of stochastic above we assume deterministic And-rules for simplicity. In AOGs, we also propose a domain-independent inference al- principle, any uncertainty in an And-rule can be equivalently gorithm and study its tractability (section 2.2). Below we start represented by a set of Or-rules each invoking a different copy with the definition of stochastic context-free AOGs, which are of the And-rule. the most basic form of stochastic AOGs and are used as the Fig. 1(a) shows an example stochastic context-free AOG of skeleton in building more advanced stochastic AOGs. line drawings. Each terminal or nonterminal node represents A stochastic context-free AOG is defined as a 5-tuple an image patch and its parameter is a 2D vector representing ⌃,N,S,✓,R : the position of the patch in the image. Each terminal node de- h i notes a line segment of a specific orientation while each non- ⌃ is a set of terminal nodes representing atomic patterns that terminal node denotes a class of line drawing patterns. The are not decomposable; start symbol S denotes a class of line drawing images (e.g., N is a set of nonterminal nodes representing high-level pat- images of animal faces). In each And-rule, the parameter re- terns, which is divided into two disjoint sets: And-nodes lation specifies the relative positions between the sub-patterns and Or-nodes; and the parameter function specifies the relative positions be- tween the composite pattern and the sub-patterns. S N is a start symbol that represents a complete pattern; 2 With a stochastic context-free AOG, one can generate a ✓ is a function that maps an instance of a terminal or nonter- compositional structure by starting from a data sample con- minal node x to a parameter ✓x (the parameter can take taining only the start symbol S and recursively applying the any form such as a vector or a complex data structure; grammar rules in R to convert nonterminal nodes in the data denote the maximal parameter size by m✓); sample until the data sample contains only terminal nodes.
Recommended publications
  • Computer Vision Stochastic Grammars for Scene Parsing
    Computer Vision Stochastic Grammars for Scene Parsing Song-Chun Zhu Ying Nian Wu August 18, 2020 Contents 0.1 This Book in the Series . ix Part I Stochastic Grammars in Vision 1 Introduction 3 1.1 Vision as Joint Parsing of Objects, Scenes, and Events . .3 1.2 Unified Representation for Models, Algorithms, and Attributes . .6 1.2.1 Three Families of Probabilistic Models . .6 1.2.2 Dynamics of Three Inferential Channels . .7 1.2.3 Three Attributes associated with each Node . .7 1.3 Missing Themes in Popular Data-Driven Approaches . .8 1.3.1 Top-Down Inference in Space and Time . .8 1.3.2 Vision is to be a Continuous Computational Process . .9 1.3.3 Resolving Ambiguities and Preserving Distinct Solutions . 11 1.3.4 Vision is Driven by a Large Number of Tasks . 12 1.4 Scope of Book: Compositional Patterns in the Middle Entropy Regime . 13 1.4.1 Information Scaling and Three Entropy Regimes . 14 1.4.2 Organization of Book . 14 2 Overview of Stochastic Grammar 17 2.1 Grammar as a Universal Representation of Intelligence . 17 2.2 An Empiricist’s View of Grammars . 18 2.3 The Formalism of Grammars . 20 2.4 The Mathematical Structure of Grammars . 21 2.5 Stochastic Grammar . 23 2.6 Ambiguity and Overlapping Reusable Parts . 24 2.7 Stochastic Grammar with Context . 26 3 Spatial And-Or Graph 29 3.1 Three New Issues in Image Grammars in Contrast to Language . 29 3.2 Visual Vocabulary . 31 3.2.1 The Hierarchical Visual Vocabulary – the "Lego Land" .
    [Show full text]
  • UNIVERSITY of CALIFORNIA Los Angeles Human Activity
    UNIVERSITY OF CALIFORNIA Los Angeles Human Activity Understanding and Prediction with Stochastic Grammar A thesis submitted in partial satisfaction of the requirements for the degree Master of Science in Computer Science by Baoxiong Jia 2019 c Copyright by Baoxiong Jia 2019 ABSTRACT OF THE THESIS Human Activity Understanding and Prediction with Stochastic Grammar by Baoxiong Jia Master of Science in Computer Science University of California, Los Angeles, 2019 Professor Song-Chun Zhu, Chair Video understanding is a booming research problem in computer vision. With its innate fea- ture where spatial and temporal information entangles with each other, video understand- ing has been challenging mainly because of the difficulty for having a unified framework where these two aspects can be modeled jointly. Among the tasks in video understand- ing, human activity understanding and prediction serve as a good starting point where the spatial-temporal reasoning capability of learning modules can be tested. Most of the current approaches towards solving the human activity understanding and prediction problems use deep neural networks for spatial-temporal reasoning. However, this type of approach lacks the ability to reason beyond the local frames and conduct long-term temporal reasoning. On the other hand, stochastic grammar models are used to model observed sequences on a symbolic level with all history information considered, but they perform poorly on handling noisy input sequences. Given these insights and problems of current approaches, we propose the generalized Earley parser for bridging the gap between sequence inputs and symbolic grammars. By combining the advantages of these two types of methods, we show that the proposed model achieves a better performance on both human activity recognition and future prediction.
    [Show full text]
  • Using an Annotated Corpus As a Stochastic Grammar
    Using an Annotated Corpus as a Stochastic Grammar Rens Bod Department of Computational Linguistics University of Amsterdam Spuistraat 134 NL-1012 VB Amsterdam [email protected] Abstract plausible parse(s) of an input string from the implausible ones. In stochastic language processing, it In Data Oriented Parsing (DOP), an annotated is assumed that the most plausible parse of an input corpus is used as a stochastic grammar. An string is its most probable parse. Most instantiations input string is parsed by combining subtrees of this idea estimate the probability of a parse by from the corpus. As a consequence, one parse assigning application probabilities to context free tree can usually be generated by several rewrite roles (Jelinek, 1990), or by assigning derivations that involve different subtrces. This combination probabilities to elementary structures leads to a statistics where the probability of a (Resnik, 1992; Schabes, 1992). parse is equal to the sum of the probabilities of all its derivations. In (Scha, 1990) an informal There is some agreement now that context free rewrite introduction to DOP is given, while (Bed, rules are not adequate for estimating the probability of 1992a) provides a formalization of the theory. a parse, since they cannot capture syntactie/lexical In this paper we compare DOP with other context, and hence cannot describe how the probability stochastic grammars in the context of Formal of syntactic structures or lexical items depends on that Language Theory. It it proved that it is not context. In stochastic tree-adjoining grammar possible to create for every DOP-model a (Schabes, 1992), this lack of context-sensitivity is strongly equivalent stochastic CFG which also overcome by assigning probabilities to larger assigns the same probabilities to the parses.
    [Show full text]
  • GRAMMAR IS GRAMMAR and USAGE IS USAGE Frederick J
    GRAMMAR IS GRAMMAR AND USAGE IS USAGE Frederick J. Newmeyer Department of Linguistics University of Washington Seattle, WA 98195-4340 [email protected] January 2003 1. Usage-based models of grammar The late 1960s was an exciting time to enter field of linguistics for someone like me, whose interests tended to syntax and semantics.1 My first year as a graduate student witnessed the birth of the approach called ‘generative semantics’. Generative semantics promised to totally revolutionize the field of transformational generative grammar, which itself was barely a decade old at the time. With each passing year, generative semanticists declared that some seemingly well-established boundary was nothing but an illusion. It all began in 1967 when George Lakoff and Haj Ross challenged the existence of the level of Deep Structure, and with it the boundary between syntax and semantics (Lakoff and Ross 1967/1976). The following year Jim McCawley argued that syntactic and lexical structures were formal objects of the same sort (McCawley 1968). Then in quick succession, the dividing line between semantics and pragmatics (Lakoff 1970/1972), grammaticality and ungrammaticality (Lakoff 1973), category membership and non-membership (Ross 1973a; b), and, finally, grammar and usage (Lakoff 1974) were all cast into doubt. At the same time, many sociolinguists were proposing models in which statistical facts about the speech community were incorporated into grammatical rules (Labov 1969; 1972). But by the late 1970s, generative semantics, for reasons I have discussed elsewhere (Newmeyer 1986), had all but disappeared. Most syntacticians had re- embraced the boundaries whose demise had been heralded only a few years earlier.
    [Show full text]
  • Application of Stochastic Grammars to Understanding Action
    Application of sto chastic grammars to understanding action by Yuri A Ivanov MS Computer Science State AcademyofAir and Space Instrumentation St Petersburg Russia February Submitted to the Program in Media Arts and Sciences Scho ol of Architecture and Planning in partial fulllmentofthe requirements for the degree of MASTER OF SCIENCE IN MEDIA ARTS AND SCIENCES at the Massachusetts Institute of Technology February c Massachusetts Institute of Technology All Rights Reserved Signature of Author Program in Media Arts and Sciences January Certied by Aaron F Bobick LG Electronics Inc Career Development AssistantProfessor of Computational Vision Program in Media Arts and Sciences Thesis Sup ervisor Accepted by Stephen A Benton Chairp erson Departmental Committee on Graduate Students Program in Media Arts and Sciences Application of sto chastic grammars to understanding action by Yuri A Ivanov Submitted to the Program in Media Arts and Sciences Scho ol of Architecture and Planning on January in partial fulllmentofthe requirements for the degree of Master of Science in Media Arts and Sciences Abstract This thesis addresses the problem of using probabilistic formal languages to describ e and understand actions with explicit structure The thesis explores several mechanisms of pars ing the uncertain input string aided byastochastic contextfree grammar This metho d originating in sp eechrecognition lets us combine a statistical recognition approach with a syntactical one in a unied syntacticsemantic framework for action recognition The basic
    [Show full text]
  • W. G. M., a Stochastic Model of Language Change Through Social
    A STOCHASTIC MODEL OF LANGUAGE CHANGE THROUGH SOCIAL STRUCTURE AND PREDICTION-DRIVEN INSTABILITY W. GARRETT MITCHENER ABSTRACT. We develop a new stochastic model of language learning and change that incorporates variable speech and age structure. Children detect correlations between age and speech, predict how the language is changing, and speak according to that prediction. Although the popu- lation tends to be dominated by one idealized grammar or another, the learning process amplifies chance correlation between age and speech, generating prediction-driven instability and spontaneous language change. The model is formulated as a Markov chain, and a stochastic differential equation. It exhibits the abrupt monotonic shifts among equilibria char- acteristic of language changes. Fundamental results are proved. 1. THE PARADOX OF LANGUAGE CHANGE One of the primary tools in the field of linguistics is the idealized gram- mar, that is, a formalism that identifies correct utterances (1, 2). Much of the research on how children acquire their native language focuses on how they might choose an idealized grammar from many innate possibilities on the basis of example sentences from the surrounding society (3–5). From the perspective of idealized grammar, language change is paradoxical: Children acquire their native language accurately and communicate with adults from preceding generations, yet over time, the language can change drastically. Some changes may be attributed to an external event, such as political up- heaval, but not every instance of language change seems to have an external cause. Despite their variability, languages do maintain considerable short- term stability, consistently accepting and rejecting large classes of sentences Key words and phrases.
    [Show full text]
  • The Plasticity of Grammar
    The Plasticity of Grammar JOAN BRESNAN Stanford University & CSLI, The Center for the Study of Language and Information The plasticity of grammar A simplifying assumption of linguistic theory: knowledge of language is characterized by a static, categorical system of grammar. The plasticity of grammar A simplifying assumption of linguistic theory: knowledge of language is characterized by a static, categorical system of grammar. —A fruitful idealization, but it ultimately underestimates human language capacities. The plasticity of grammar A simplifying assumption of linguistic theory: knowledge of language is characterized by a static, categorical system of grammar. —A fruitful idealization, but it ultimately underestimates human language capacities. Language users can match the probabilities of linguistic features of the environment and they have powerful predictive capabilities that enable them to anticipate the variable linguistic choices of others. The plasticity of grammar A simplifying assumption of linguistic theory: knowledge of language is characterized by a static, categorical system of grammar. —A fruitful idealization, but it ultimately underestimates human language capacities. Language users can match the probabilities of linguistic features of the environment and they have powerful predictive capabilities that enable them to anticipate the variable linguistic choices of others. Therefore, a strongly contrasting hypothesis: grammar itself is inherently variable and stochastic in nature, rather than categorical and algebraic. Implicit knowledge of variability Low-level continuous phonetic variation is sensitive to high-level construction probabilities in speech production (Gahl & Garnsey 2004, Tily et al. 2009, Kuperman & Bresnan 2012) Construction probability is not merely a summary statistic for determinative individual factors affecting pronunciation, but has a direct effect in itself (Kuperman & Bresnan 2012) Illustrations from Victor Kuperman & Joan Bresnan (2012).
    [Show full text]
  • Calibrating Generative Models: the Probabilistic Chomsky-Schutzenberger¨ Hierarchy∗
    Calibrating Generative Models: The Probabilistic Chomsky-Schutzenberger¨ Hierarchy∗ Thomas F. Icard Stanford University [email protected] December 14, 2019 1 Introduction and Motivation Probabilistic computation enjoys a central place in psychological modeling and theory, ranging in application from perception (Knill and Pouget, 2004; Orban´ et al., 2016), at- tention (Vul et al., 2009), and decision making (Busemeyer and Diederich, 2002; Wilson et al., 2014) to the analysis of memory (Ratcliff, 1978; Abbott et al., 2015), categorization (Nosofsky and Palmeri, 1997; Sanborn et al., 2010), causal inference (Denison et al., 2013), and various aspects of language processing (Chater and Manning, 2006). Indeed, much of cognition and behavior is usefully modeled as stochastic, whether this is ultimately due to indeterminacy at the neural level or at some higher level (see, e.g., Glimcher 2005). Ran- dom behavior may even enhance an agent’s cognitive efficiency (see, e.g., Icard 2019). A probabilistic process can be described in two ways. On one hand we can specify a stochastic procedure, which may only implicitly define a distribution on possible out- puts of the process. Familiar examples include well known classes of generative mod- els such as finite Markov processes and probabilistic automata (Miller, 1952; Paz, 1971), Boltzmann machines (Hinton and Sejnowski, 1983), Bayesian networks (Pearl, 1988), topic models (Griffiths et al., 2007), and probabilistic programming languages (Tenenbaum et al., 2011; Goodman and Tenenbaum, 2016). On the other hand we can specify an analytical ex- pression explicitly describing the distribution over outputs. Both styles of description are commonly employed, and it is natural to ask about the relationship between them.
    [Show full text]
  • Stochastic Definite Clause Grammars
    Stochastic Definite Clause Grammars Christian Theil Have Research group PLIS: Programming, Logic and Intelligent Systems Department of Communication, Business and Information Technologies Roskilde University, P.O.Box 260, DK-4000 Roskilde, Denmark [email protected] Abstract where corpus annotations dictate the derivations, can This paper introduces Stochastic Definite Clause be done by counting expansions used in the annota- Grammars, a stochastic variant of the well- tions. Estimation with incomplete data can be ac- known Definite Clause Grammars. The grammar complished using the Expectation-Maximization (EM) formalism supports parameter learning from an- algorithm [8]. notated or unannotated corpora and provides a In stochastic unification grammars, the choice of mechanism for parse selection by means of sta- rules to expand is stochastic and the values assigned tistical inference. Unlike probabilistic context- free grammars, it is a context-sensitive gram- to unification variables are determined implicitly by mar formalism and it has the ability to model rule selection. This means that in some derivations, cross-serial dependencies in natural language. instances of the same logic variable may get different SDCG also provides some syntax extensions values and unification will fail as result. which makes it possible to write more compact Some of the first attempts to define stochastic uni- grammars and makes it straight-forward to add fication grammars did not address the issue of how lexicalization schemes to a grammar. they should be trained. Brew [2] and Eisele [9] tries to address this problem using EM, but their methods have problems handling cases where variables fails to 1 Introduction and background unify.
    [Show full text]
  • Stochastic Attribute-Value Grammars
    Stochastic Attribute-Value Grammars Steven P. Abney* AT&T Laboratories Probabilistic analogues of regular and context-free grammars are well known in computational linguistics, and currently the subject of intensive research. To date, however, no satisfactory probabilistic analogue of attribute-value grammars has been proposed: previous attempts have failed to define an adequate parameter-estimation algorithm. In the present paper, I define stochastic attribute-value grammars and give an algorithm for computing the maximum-likelihood estimate of their parameters. The estimation algorithm is adapted from Della Pietra, Della Pietra, and Lafferty (1995). To estimate model parameters, it is necessary to compute the expectations of certain functions under random fields. In the appli- cation discussed by Della Pietra, Della Pietra, and Lafferty (representing English orthographic constraints), Gibbs sampling can be used to estimate the needed expectations. The fact that attribute-value grammars generate constrained languages makes Gibbs sampling inapplicable, but I show that sampling can be done using the more general Metropolis-Hastings algorithm. 1. Introduction Stochastic versions of regular grammars and context-free grammars have received a great deal of attention in computational linguistics for the last several years, and ba- sic techniques of stochastic parsing and parameter estimation have been known for decades. However, regular and context-free grammars are widely deemed linguisti- cally inadequate; standard grammars in computational linguistics are attribute-value (AV) grammars of some variety. Before the advent of statistical methods, regular and context-free grammars were considered too inexpressive for serious consideration, and even now the reliance on stochastic versions of the less-expressive grammars is often seen as an expedient necessitated by the lack of an adequate stochastic version of attribute-value grammars.
    [Show full text]
  • Unsupervised Language Acquisition: Theory and Practice
    Unsupervised Language Acquisition: Theory and Practice Alexander Simon Clark Submitted for the degree of D.Phil. University of Sussex September, 2001 Declaration I hereby declare that this thesis has not been submitted, either in the same or different form, to this or any other university for a degree. Signature: Acknowledgements First, I would like to thank Bill Keller, for his supervision over the past three years. I would like to thank my colleagues at ISSCO for making me welcome and especially Susan Armstrong for giving me the job in the first place, and various people for helpful comments and discussions including Chris Manning, Dan Klein, Eric Gaussier, Nicola Cancedda, Franck Thollard, Andrei Popescu-Beilis, Menno van Zaanen and numerous other people including Sonia Halimi for check- ing my Arabic. I would also like to thank Gerald Gazdar and Geoffrey Sampson for their helpful comments as part of my thesis committee. I would also like to thank all of the people that have worked on the various software packages and operating systems that I have used, such as LaTeX, Gnu/Linux, and the gcc compiler, as well as the people who have worked on the preparation of the various corpora I have used: in particular I am grateful to John McCarthy and Ramin Nakisa for allowing me to use their painstakingly collected data sets for Arabic. Finally I would like to thank my wife, Dr Clare Hornsby, for her unfailing support and encouragement (in spite of her complete lack of understanding of the subject) and my children, Lily and Francis, for putting up with my absences both physical and mental, and showing me how language acquisition is really done – over-regularisation of past tense verbs included.
    [Show full text]
  • A Stochastic Grammar of Images
    A Stochastic Grammar of Images The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Zhu, Song Chun, and David Bryant Mumford. 2006. A stochastic grammar of images. Foundations and Trends in Computer Graphics and Vision 2(4): 259-362. Published Version doi:10.1561/0600000018 Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:3637153 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA Foundations and TrendsR in Computer Graphics and Vision Vol. 2, No. 4 (2006) 259–362 c 2007 S.-C. Zhu and D. Mumford DOI: 10.1561/0600000018 A Stochastic Grammar of Images Song-Chun Zhu1,∗ and David Mumford2 1 University of California, Los Angeles, USA, [email protected] 2 Brown University, USA, David [email protected] Abstract This exploratory paper quests for a stochastic and context sensitive grammar of images. The grammar should achieve the following four objectives and thus serves as a unified framework of representation, learning, and recognition for a large number of object categories. (i) The grammar represents both the hierarchical decompositions from scenes, to objects, parts, primitives and pixels by terminal and non-terminal nodes and the contexts for spatial and functional relations by horizon- tal links between the nodes. It formulates each object category as the set of all possible valid configurations produced by the grammar.
    [Show full text]