Learning Dependency-Based Compositional Semantics

Total Page:16

File Type:pdf, Size:1020Kb

Learning Dependency-Based Compositional Semantics Learning Dependency-Based Compositional Semantics Percy Liang∗ University of California, Berkeley Michael I. Jordan∗∗ University of California, Berkeley Dan Klein† University of California, Berkeley Suppose we want to build a system that answers a natural language question by representing its semantics as a logical form and computing the answer given a structured database of facts. The core part of such a system is the semantic parser that maps questions to logical forms. Semantic parsers are typically trained from examples of questions annotated with their target logical forms, but this type of annotation is expensive. Our goal is to instead learn a semantic parser from question–answer pairs, where the logical form is modeled as a latent variable. We develop a new semantic formalism, dependency-based compositional semantics (DCS) and define a log-linear distribution over DCS logical forms. The model parameters are estimated using a simple procedure that alternates between beam search and numerical optimization. On two standard semantic parsing benchmarks, we show that our system obtains comparable accuracies to even state-of-the-art systems that do require annotated logical forms. 1. Introduction One of the major challenges in natural language processing (NLP) is building systems that both handle complex linguistic phenomena and require minimal human effort. The difficulty of achieving both criteria is particularly evident in training semantic parsers, where annotating linguistic expressions with their associated logical forms is expensive but until recently, seemingly unavoidable. Advances in learning latent-variable models, however, have made it possible to progressively reduce the amount of supervision ∗ Computer Science Division, University of California, Berkeley, CA 94720, USA. E-mail: [email protected]. ∗∗ Computer Science Division and Department of Statistics, University of California, Berkeley, CA 94720, USA. E-mail: [email protected]. † Computer Science Division, University of California, Berkeley, CA 94720, USA. E-mail: [email protected]. Submission received: 12 September 2011; revised submission received: 19 February 2012; accepted for publication: 18 April 2012. doi:10.1162/COLI a 00127 No rights reserved. This work was authored as part of the Contributor’s official duties as an Employee of the United States Government and is therefore a work of the United States Government. In accordance with 17 U.S.C. 105, no copyright protection is available for such works under U.S. law. Computational Linguistics Volume 39, Number 2 required for various semantics-related tasks (Zettlemoyer and Collins 2005; Branavan et al. 2009; Liang, Jordan, and Klein 2009; Clarke et al. 2010; Artzi and Zettlemoyer 2011; Goldwasser et al. 2011). In this article, we develop new techniques to learn accurate semantic parsers from even weaker supervision. We demonstrate our techniques on the concrete task of building a system to answer questions given a structured database of facts; see Figure 1 for an example in the domain of U.S. geography. This problem of building natural language interfaces to databases (NLIDBs) has a long history in NLP, starting from the early days of artificial intelligence with systems such as LUNAR (Woods, Kaplan, and Webber 1972), CHAT-80 (Warren and Pereira 1982), and many others (see Androutsopoulos, Ritchie, and Thanisch [1995] for an overview). We believe NLIDBs provide an appropriate starting point for semantic parsing because they lead directly to practical systems, and they allowus to temporarily sidestep intractable philosophical questions on howto represent meaning in general. Early NLIDBs were quite successful in their respective limited domains, but because these systems were constructed from manually built rules, they became difficult to scale up, both to other domains and to more complex utterances. In response, against the backdrop of a statistical revolution in NLP during the 1990s, researchers began to build systems that could learn from examples, with the hope of overcoming the limitations of rule-based methods. One of the earliest statistical efforts was the CHILL system (Zelle and Mooney 1996), which learned a shift-reduce semantic parser. Since then, there has been a healthy line of work yielding increasingly more accurate semantic parsers by using newsemantic representations and machine learning techniques (Miller et al. 1996; Zelle and Mooney 1996; Tang and Mooney 2001; Ge and Mooney 2005; Kate, Wong, and Mooney 2005; Zettlemoyer and Collins 2005; Kate and Mooney 2006; Wong and Mooney 2006; Kate and Mooney 2007; Wong and Mooney 2007; Zettlemoyer and Collins 2007; Kwiatkowski et al. 2010, 2011). Although statistical methods provided advantages such as robustness and portabil- ity, however, their application in semantic parsing achieved only limited success. One of the main obstacles was that these methods depended crucially on having examples of utterances paired with logical forms, and this requires substantial human effort to obtain. Furthermore, the annotators must be proficient in some formal language, which drastically reduces the size of the annotator pool, dampening any hope of acquiring enough data to fulfill the vision of learning highly accurate systems. In response to these concerns, researchers have recently begun to explore the pos- sibility of learning a semantic parser without any annotated logical forms (Clarke et al. Figure 1 The concrete objective: A system that answers natural language questions given a structured database of facts. An example is shown in the domain of U.S. geography. 390 Liang, Jordan, and Klein Learning Dependency-Based Compositional Semantics Figure 2 Our statistical methodology consists of two steps: (i) semantic parsing (p(z | x; θ)): an utterance x is mapped to a logical form z by drawing from a log-linear distribution parametrized by a vector θ; and (ii) evaluation ([[z]]w): the logical form z is evaluated with respect to the world w (database of facts) to deterministically produce an answer y. The figure also shows an example configuration of the variables around the graphical model. Logical forms z are represented as labeled trees. During learning, we are given w and (x, y)pairs(shadednodes)andtrytoinfer the latent logical forms z and parameters θ. 2010; Artzi and Zettlemoyer 2011; Goldwasser et al. 2011; Liang, Jordan, and Klein 2011). It is in this vein that we develop our present work. Specifically, given a set of (x, y) example pairs, where x is an utterance (e.g., a question) and y is the corresponding answer, we wish to learn a mapping from x to y. What makes this mapping particularly interesting is that it passes through a latent logical form z, which is necessary to capture the semantic complexities of natural language. Also note that whereas the logical form z was the end goal in much of earlier work on semantic parsing, for us it is just an intermediate variable—a means towards an end. Figure 2 shows the graphical model which captures the learning setting we just described: The question x, answer y,and world/database w are all observed. We want to infer the logical forms z and the parameters θ of the semantic parser, which are unknown quantities. Although liberating ourselves from annotated logical forms reduces cost, it does increase the difficulty of the learning problem. The core challenge here is program induction: On each example (x, y), we need to efficiently search over the exponential space of possible logical forms (programs) z and find ones that produce the target answer y, a computationally daunting task. There is also a statistical challenge: How do we parametrize the mapping from utterance x to logical form z so that it can be learned from only the indirect signal y? To address these two challenges, we must first discuss the issue of semantic representation. There are two basic questions here: (i) what 391 Computational Linguistics Volume 39, Number 2 should the formal language for the logical forms z be, and (ii) what are the compositional mechanisms for constructing those logical forms? The semantic parsing literature has considered many different formal languages for representing logical forms, including SQL (Giordani and Moschitti 2009), Prolog (Zelle and Mooney 1996; Tang and Mooney 2001), a simple functional query language called FunQL (Kate, Wong, and Mooney 2005), and lambda calculus (Zettlemoyer and Collins 2005), just to name a few. The construction mechanisms are equally diverse, in- cluding synchronous grammars (Wong and Mooney 2007), hybrid trees (Lu et al. 2008), Combinatory Categorial Grammars (CCG) (Zettlemoyer and Collins 2005), and shift- reduce derivations (Zelle and Mooney 1996). It is worth pointing out that the choice of formal language and the construction mechanism are decisions which are really more orthogonal than is often assumed—the former is concerned with what the logical forms look like; the latter, with how to generate a set of possible logical forms compositionally given an utterance. (Howto score these logical forms is yet another dimension.) Existing systems are rarely based on the joint design of the formal language and the construction mechanism; one or the other is often chosen for convenience from existing implementations. For example, Prolog and SQL have often been chosen as formal languages for convenience in end applications, but they were not designed for representing the semantics of natural language, and, as a result, the construction mechanism that bridges the gap between natural
Recommended publications
  • UNIT-I Mathematical Logic Statements and Notations
    UNIT-I Mathematical Logic Statements and notations: A proposition or statement is a declarative sentence that is either true or false (but not both). For instance, the following are propositions: “Paris is in France” (true), “London is in Denmark” (false), “2 < 4” (true), “4 = 7 (false)”. However the following are not propositions: “what is your name?” (this is a question), “do your homework” (this is a command), “this sentence is false” (neither true nor false), “x is an even number” (it depends on what x represents), “Socrates” (it is not even a sentence). The truth or falsehood of a proposition is called its truth value. Connectives: Connectives are used for making compound propositions. The main ones are the following (p and q represent given propositions): Name Represented Meaning Negation ¬p “not p” Conjunction p q “p and q” Disjunction p q “p or q (or both)” Exclusive Or p q “either p or q, but not both” Implication p ⊕ q “if p then q” Biconditional p q “p if and only if q” Truth Tables: Logical identity Logical identity is an operation on one logical value, typically the value of a proposition that produces a value of true if its operand is true and a value of false if its operand is false. The truth table for the logical identity operator is as follows: Logical Identity p p T T F F Logical negation Logical negation is an operation on one logical value, typically the value of a proposition that produces a value of true if its operand is false and a value of false if its operand is true.
    [Show full text]
  • Chapter 1 Logic and Set Theory
    Chapter 1 Logic and Set Theory To criticize mathematics for its abstraction is to miss the point entirely. Abstraction is what makes mathematics work. If you concentrate too closely on too limited an application of a mathematical idea, you rob the mathematician of his most important tools: analogy, generality, and simplicity. – Ian Stewart Does God play dice? The mathematics of chaos In mathematics, a proof is a demonstration that, assuming certain axioms, some statement is necessarily true. That is, a proof is a logical argument, not an empir- ical one. One must demonstrate that a proposition is true in all cases before it is considered a theorem of mathematics. An unproven proposition for which there is some sort of empirical evidence is known as a conjecture. Mathematical logic is the framework upon which rigorous proofs are built. It is the study of the principles and criteria of valid inference and demonstrations. Logicians have analyzed set theory in great details, formulating a collection of axioms that affords a broad enough and strong enough foundation to mathematical reasoning. The standard form of axiomatic set theory is denoted ZFC and it consists of the Zermelo-Fraenkel (ZF) axioms combined with the axiom of choice (C). Each of the axioms included in this theory expresses a property of sets that is widely accepted by mathematicians. It is unfortunately true that careless use of set theory can lead to contradictions. Avoiding such contradictions was one of the original motivations for the axiomatization of set theory. 1 2 CHAPTER 1. LOGIC AND SET THEORY A rigorous analysis of set theory belongs to the foundations of mathematics and mathematical logic.
    [Show full text]
  • Folktale Documentation Release 1.0
    Folktale Documentation Release 1.0 Quildreen Motta Nov 05, 2017 Contents 1 Guides 3 2 Indices and tables 5 3 Other resources 7 3.1 Getting started..............................................7 3.2 Folktale by Example........................................... 10 3.3 API Reference.............................................. 12 3.4 How do I................................................... 63 3.5 Glossary................................................. 63 Python Module Index 65 i ii Folktale Documentation, Release 1.0 Folktale is a suite of libraries for generic functional programming in JavaScript that allows you to write elegant modular applications with fewer bugs, and more reuse. Contents 1 Folktale Documentation, Release 1.0 2 Contents CHAPTER 1 Guides • Getting Started A series of quick tutorials to get you up and running quickly with the Folktale libraries. • API reference A quick reference of Folktale’s libraries, including usage examples and cross-references. 3 Folktale Documentation, Release 1.0 4 Chapter 1. Guides CHAPTER 2 Indices and tables • Global Module Index Quick access to all modules. • General Index All functions, classes, terms, sections. • Search page Search this documentation. 5 Folktale Documentation, Release 1.0 6 Chapter 2. Indices and tables CHAPTER 3 Other resources • Licence information 3.1 Getting started This guide will cover everything you need to start using the Folktale project right away, from giving you a brief overview of the project, to installing it, to creating a simple example. Once you get the hang of things, the Folktale By Example guide should help you understanding the concepts behind the library, and mapping them to real use cases. 3.1.1 So, what’s Folktale anyways? Folktale is a suite of libraries for allowing a particular style of functional programming in JavaScript.
    [Show full text]
  • Theorem Proving in Classical Logic
    MEng Individual Project Imperial College London Department of Computing Theorem Proving in Classical Logic Supervisor: Dr. Steffen van Bakel Author: David Davies Second Marker: Dr. Nicolas Wu June 16, 2021 Abstract It is well known that functional programming and logic are deeply intertwined. This has led to many systems capable of expressing both propositional and first order logic, that also operate as well-typed programs. What currently ties popular theorem provers together is their basis in intuitionistic logic, where one cannot prove the law of the excluded middle, ‘A A’ – that any proposition is either true or false. In classical logic this notion is provable, and the_: corresponding programs turn out to be those with control operators. In this report, we explore and expand upon the research about calculi that correspond with classical logic; and the problems that occur for those relating to first order logic. To see how these calculi behave in practice, we develop and implement functional languages for propositional and first order logic, expressing classical calculi in the setting of a theorem prover, much like Agda and Coq. In the first order language, users are able to define inductive data and record types; importantly, they are able to write computable programs that have a correspondence with classical propositions. Acknowledgements I would like to thank Steffen van Bakel, my supervisor, for his support throughout this project and helping find a topic of study based on my interests, for which I am incredibly grateful. His insight and advice have been invaluable. I would also like to thank my second marker, Nicolas Wu, for introducing me to the world of dependent types, and suggesting useful resources that have aided me greatly during this report.
    [Show full text]
  • Logical Vs. Natural Language Conjunctions in Czech: a Comparative Study
    ITAT 2016 Proceedings, CEUR Workshop Proceedings Vol. 1649, pp. 68–73 http://ceur-ws.org/Vol-1649, Series ISSN 1613-0073, c 2016 K. Prikrylová,ˇ V. Kubon,ˇ K. Veselovská Logical vs. Natural Language Conjunctions in Czech: A Comparative Study Katrin Prikrylová,ˇ Vladislav Kubon,ˇ and Katerinaˇ Veselovská Charles University in Prague, Faculty of Mathematics and Physics Czech Republic {prikrylova,vk,veselovska}@ufal.mff.cuni.cz Abstract: This paper studies the relationship between con- ceptions and irregularities which do not abide the rules as junctions in a natural language (Czech) and their logical strictly as it is the case in logic. counterparts. It shows that the process of transformation Primarily due to this difference, the transformation of of a natural language expression into its logical representa- natural language sentences into their logical representation tion is not straightforward. The paper concentrates on the constitutes a complex issue. As we are going to show in most frequently used logical conjunctions, and , and it the subsequent sections, there are no simple rules which analyzes the natural language phenomena which∧ influence∨ would allow automation of the process – the majority of their transformation into logical conjunction and disjunc- problematic cases requires an individual approach. tion. The phenomena discussed in the paper are temporal In the following text we are going to restrict our ob- sequence, expressions describing mutual relationship and servations to the two most frequently used conjunctions, the consequences of using plural. namely a (and) and nebo (or). 1 Introduction and motivation 2 Sentences containing the conjunction a (and) The endeavor to express natural language sentences in the form of logical expressions is probably as old as logic it- The initial assumption about complex sentences contain- self.
    [Show full text]
  • Presupposition Projection and Entailment Relations
    Presupposition Projection and Entailment Relations Amaia Garcia Odon i Acknowledgements I would like to thank the members of my thesis committee, Prof. Dr. Rob van der Sandt, Dr. Henk Zeevat, Dr. Isidora Stojanovic, Dr. Cornelia Ebert and Dr. Enric Vallduví for having accepted to be on my thesis committee. I am extremely grateful to my adviser, Louise McNally, and to Rob van der Sandt. Without their invaluable help, I would not have written this dissertation. Louise has been a wonderful adviser, who has always provided me with excellent guid- ance and continuous encouragement. Rob has been a mentor who has generously spent much time with me, teaching me Logic, discussing my work, and helping me clear up my thoughts. Very special thanks to Henk Zeevat for having shared his insights with me and for having convinced me that it was possible to finish on time. Thanks to Bart Geurts, Noor van Leusen, Sammie Tarenskeen, Bob van Tiel, Natalia Zevakhina and Corien Bary in Nijmegen; to Paul Dekker, Jeroen Groe- nendijk, Frank Veltman, Floris Roelofsen, Morgan Mameni, Raquel Fernández and Margot Colinet in Amsterdam; to Fernando García Murga, Agustín Vicente, Myriam Uribe-Etxebarria, Javier Ormazabal, Vidal Valmala, Gorka Elordieta, Urtzi Etxeberria and Javi Fernández in Vitoria-Gasteiz, to David Beaver and Mari- bel Romero. Also thanks to the people I met at the ESSLLIs of Bordeaux, Copen- hagen and Ljubljana, especially to Nick Asher, Craige Roberts, Judith Tonhauser, Fenghui Zhang, Mingya Liu, Alexandra Spalek, Cornelia Ebert and Elena Pa- ducheva. I gratefully acknowledge the financial help provided by the Basque Government – Departamento de Educación, Universidades e Investigación (BFI07.96) and Fun- dación ICREA (via an ICREA Academia award to Louise McNally).
    [Show full text]
  • Introduction to Logic 1
    Introduction to logic 1. Logic and Artificial Intelligence: some historical remarks One of the aims of AI is to reproduce human features by means of a computer system. One of these features is reasoning. Reasoning can be viewed as the process of having a knowledge base (KB) and manipulating it to create new knowledge. Reasoning can be seen as comprised of 3 processes: 1. Perceiving stimuli from the environment; 2. Translating the stimuli in components of the KB; 3. Working on the KB to increase it. We are going to deal with how to represent information in the KB and how to reason about it. We use logic as a device to pursue this aim. These notes are an introduction to modern logic, whose origin can be found in George Boole’s and Gottlob Frege’s works in the XIX century. However, logic in general traces back to ancient Greek philosophers that studied under which conditions an argument is valid. An argument is any set of statements – explicit or implicit – one of which is the conclusion (the statement being defended) and the others are the premises (the statements providing the defense). The relationship between the conclusion and the premises is such that the conclusion follows from the premises (Lepore 2003). Modern logic is a powerful tool to represent and reason on knowledge and AI has traditionally adopted it as working tool. Within logic, one research tradition that strongly influenced AI is that started by the German philosopher and mathematician Gottfried Leibniz. His aim was to formulate a Lingua Rationalis to create a universal language based only on rationality, which could solve all the controversies between different parties.
    [Show full text]
  • The Composition of Logical Constants
    the morphosemantic makeup of exclusive-disjunctive markers ∗ Moreno Mitrovi´c university of cyprus & bled institute . abstract. The spirit of this paper is strongly decompositional and its aim to meditate on the idea that natural language conjunction and disjunction markers do not necessarily (directly) incarnate Boo- lean terms like ‘0’ and ‘1’,respectively. Drawing from a rich collec- tion of (mostly dead) languages (Ancient Anatolian, Homeric Greek, Tocharian, Old Church Slavonic, and North-East Caucasian), I will examine the morphosemantics of ‘XOR’ (‘exclusive or’) markers, i.e. exclusive (or strong/enriched) disjunction markers of the “ei- Do not cite without consultation. ⭒ ther ...or ...”-type, and demonstrate that the morphology of the XOR marker does not only contain the true (generalised) disjunc- tion marker (I dub it κ), as one would expect on the null (Boolean hy- pothesis), but that the XOR-marker also contains the (generalised) conjunction marker (I dub it μ). After making the case for a fine- April 15, 2017 structure of the Junction Phrase (JP), a common structural denom- ⭒ inator for con- and dis-junction, the paper proposes a new syntax for XOR constructions involving five functional heads (two pairs of κ ver. 6 and μ markers, forming the XOR-word and combining with the respective coordinand, and a J-head pairing up the coordinands). draft I then move on to compose the semantics of the syntactically de- ⭒ composed structure by providing a compositional account obtain- ing the exclusive component, qua the semantic/pragmatic signa- ture of these markers. To do so, I rely on an exhaustification-based system of ‘grammaticised implicatures’ (Chierchia, 2013b) in as- manuscript suming silent exhaustification operators in the narrow syntax, which (in concert with the presence of alternative-triggering κ- and μ-ope- rators) trigger local exclusive (scalar) implicature (SI) computation.
    [Show full text]
  • Working with Logic
    Working with Logic Jordan Paschke September 2, 2015 One of the hardest things to learn when you begin to write proofs is how to keep track of everything that is going on in a theorem. Using a proof by contradiction, or proving the contrapositive of a proposition, is sometimes the easiest way to move forward. However negating many different hypotheses all at once can be tricky. This worksheet is designed to help you adjust between thinking in “English” and thinking in “math.” It should hopefully serve as a reference sheet where you can quickly look up how the various logical operators and symbols interact with each other. Logical Connectives Let us consider the two propositions: P =“Itisraining.” Q =“Iamindoors.” The following are called logical connectives,andtheyarethemostcommonwaysof combining two or more propositions (or only one, in the case of negation) into a new one. Negation (NOT):ThelogicalnegationofP would read as • ( P )=“Itisnotraining.” ¬ Conjunction (AND):ThelogicalconjunctionofP and Q would be read as • (P Q) = “It is raining and I am indoors.” ∧ Disjunction (OR):ThelogicaldisjunctionofP and Q would be read as • (P Q)=“ItisrainingorIamindoors.” ∨ Implication (IF...THEN):ThelogicalimplicationofP and Q would be read as • (P = Q)=“Ifitisraining,thenIamindoors.” ⇒ 1 Biconditional (IF AND ONLY IF):ThelogicalbiconditionalofP and Q would • be read as (P Q)=“ItisrainingifandonlyifIamindoors.” ⇐⇒ Along with the implication (P = Q), there are three other related conditional state- ments worth mentioning: ⇒ Converse:Thelogicalconverseof(P = Q)wouldbereadas • ⇒ (Q = P )=“IfIamindoors,thenitisraining.” ⇒ Inverse:Thelogicalinverseof(P = Q)wouldbereadas • ⇒ ( P = Q)=“Ifitisnotraining,thenIamnotindoors.” ¬ ⇒¬ Contrapositive:Thelogicalcontrapositionof(P = Q)wouldbereadas • ⇒ ( Q = P )=“IfIamnotindoors,thenIitisnotraining.” ¬ ⇒¬ It is worth mentioning that the implication (P = Q)andthecontrapositive( Q = P ) are logically equivalent.
    [Show full text]
  • Datafun: a Functional Datalog
    Datafun: a Functional Datalog Michael Arntzenius Neelakantan R. Krishnaswami University of Birmingham (UK) {daekharel, neelakantan.krishnaswami}@gmail.com Abstract Since proof search is in general undecidable, designers of logic Datalog may be considered either an unusually powerful query lan- programming languages must be careful both about the kinds guage or a carefully limited logic programming language. Datalog of formulas they admit as programs, and about the proof search is declarative, expressive, and optimizable, and has been applied algorithm they implement. Prolog offers a very expressive language successfully in a wide variety of problem domains. However, most — full Horn clauses — and so faces an undecidable proof search use-cases require extending Datalog in an application-specific man- problem. Therefore, Prolog specifies its proof search strategy: depth- ner. In this paper we define Datafun, an analogue of Datalog sup- first goal-directed/top-down search. This lets Prolog programmers porting higher-order functional programming. The key idea is to reason about the behaviour of their programs; however, it also means track monotonicity with types. many logically natural programs fail to terminate. Notoriously, transitive closure calculations are much less elegant in Prolog Categories and Subject Descriptors F.3.2 [Logics and Meanings than one might hope, since their most natural specification is best of Programs]: Semantics of Programming Languages computed with a bottom-up (aka “forwards chaining”) proof search strategy. Keywords Prolog, Datalog, logic programming, functional pro- This view of Prolog suggests other possible design choices, gramming, domain-specific languages, type theory, denotational such as restricting the logical language so as to make proof search semantics, operational semantics, adjoint logic decidable.
    [Show full text]
  • Exclusive Or from Wikipedia, the Free Encyclopedia
    New features Log in / create account Article Discussion Read Edit View history Exclusive or From Wikipedia, the free encyclopedia "XOR" redirects here. For other uses, see XOR (disambiguation), XOR gate. Navigation "Either or" redirects here. For Kierkegaard's philosophical work, see Either/Or. Main page The logical operation exclusive disjunction, also called exclusive or (symbolized XOR, EOR, Contents EXOR, ⊻ or ⊕, pronounced either / ks / or /z /), is a type of logical disjunction on two Featured content operands that results in a value of true if exactly one of the operands has a value of true.[1] A Current events simple way to state this is "one or the other but not both." Random article Donate Put differently, exclusive disjunction is a logical operation on two logical values, typically the values of two propositions, that produces a value of true only in cases where the truth value of the operands differ. Interaction Contents About Wikipedia Venn diagram of Community portal 1 Truth table Recent changes 2 Equivalencies, elimination, and introduction but not is Contact Wikipedia 3 Relation to modern algebra Help 4 Exclusive “or” in natural language 5 Alternative symbols Toolbox 6 Properties 6.1 Associativity and commutativity What links here 6.2 Other properties Related changes 7 Computer science Upload file 7.1 Bitwise operation Special pages 8 See also Permanent link 9 Notes Cite this page 10 External links 4, 2010 November Print/export Truth table on [edit] archived The truth table of (also written as or ) is as follows: Venn diagram of Create a book 08-17094 Download as PDF No.
    [Show full text]
  • Signal Temporal Logic Meets Hamilton-Jacobi Reachability: Connections and Applications
    Signal Temporal Logic meets Hamilton-Jacobi Reachability: Connections and Applications Mo Chen1, Qizhan Tam1, Scott C. Livingston2, and Marco Pavone1? 1 Dept. of Aeronautics and Astronautics, Stanford University {mochen72, qtam, pavone}@stanford.edu 2 rerobots, Inc. [email protected] Abstract. Signal temporal logic (STL) and Hamilton-Jacobi (HJ) reachability analysis are effective mathematical tools for formally analyzing the behavior of robotic systems. STL is a specification language that uses a combination of logic and temporal operators to precisely express real-valued and time-dependent re- quirements on system behaviors. While recursively defined STL specifications are extremely expressive and controller synthesis methods exist, so far there has not been work that quantifies the set of states from which STL formulas can be satisfied. HJ reachability, on the other hand, is a method for computing the reachable set, that is the set of states from which a system is able to reach a goal while satisfying state and control constraints. While reasoning about sys- tem requirements through sets of states is useful for predetermining whether it is possible to satisfy desired system properties as well as obtaining state feedback controllers, so far the applicability of HJ reachability has been limited to relatively simple reach-avoid specifications. In this paper, we merge STL and HJ reachabil- ity into a single framework that combines the key advantage of both methods – expressiveness of specifications and set quantification. To do this, we establish a correspondence between temporal and reachability operators, and utilize the idea of least-restrictive feasible controller sets (LRFCSs) to break down controller syn- thesis for complex STL formulas into a sequence of reachability and elementary set operations.
    [Show full text]