SEMANTIC ANALYSIS of NATURAL LANGUAGE and DEFINITE CLAUSE GRAMMAR USING STATISTICAL PARSING and THESAURI By

Total Page:16

File Type:pdf, Size:1020Kb

SEMANTIC ANALYSIS of NATURAL LANGUAGE and DEFINITE CLAUSE GRAMMAR USING STATISTICAL PARSING and THESAURI By SEMANTIC ANALYSIS OF NATURAL LANGUAGE AND DEFINITE CLAUSE GRAMMAR USING STATISTICAL PARSING AND THESAURI by Björn Dagerman A thesis submitted in partial fulfillment of the requirements for the degree of BACHELOR OF SCIENCE in Computer Science Examiner: Baran Çürüklü Supervisor: Batu Akan MÄLARDALEN UNIVERSITY School of Innovation, Design and Engineering 2013 i ABSTRACT Services that rely on the semantic computations of users’ natural linguistic inputs are becoming more frequent. Computing semantic relatedness between texts is problematic due to the inherit ambiguity of natural language. The purpose of this thesis was to show how a sentence could be compared to a predefined semantic Definite Clause Grammar (DCG). Furthermore, it should show how a DCG-based system could benefit from such capabilities. Our approach combines openly available specialized NLP frameworks for statistical parsing, part-of-speech tagging and word-sense disambiguation. We compute the seman- tic relatedness using a large lexical and conceptual-semantic thesaurus. Also, we extend an existing programming language for multimodal interfaces, which uses static predefined DCGs: COactive Language Definition (COLD). That is, every word that should be accept- able by COLD needs to be explicitly defined. By applying our solution, we show how our approach can remove dependencies on word definitions and improve grammar definitions in DCG-based systems. (27 pages) ii SAMMANFATTNING Tjänster som beror på semantiska beräkningar av användares naturliga tal blir allt van- ligare. Beräkning av semantisk likhet mellan texter är problematiskt på grund av naturligt tals medförda tvetydighet. Syftet med det här examensarbetet är att visa hur en mening can jämföras med en fördefinierad semantisk Definite Clause Grammar (DCG). Dessutom bör arbetet visa hur ett DCG-baserat system kan dra nytta av den möjligheten. Vårt tillvägagångssätt kombinerar öppet tillgängliga specialiserade NLP frameworks för statistical parsing, part-of-speech tagging och worse-sense disambiguation. Vi beräknar den semantiska likheten med hjälp av en stor lexikografisk och konceptuellt semantisk synony- mordbok. Vidare utökar vi ett befintligt programmeringsspråk för multimodala gränssnitt som använder statiskt fördefinierade DCGs: COactive Language Definition (COLD). Dvs. alla ord som ska kunna accepteras av COLD måste explicit definieras. Genom att tillämpa vår lösning visar vi hur vår metod kan minska beroenden på ord-definitioner och förbättra grammatik-definitioner i DCG-baserade system. (27 sidor) iii CONTENTS Page ABSTRACT ...................................................... i SAMMANFATTNING .............................................. ii LIST OF FIGURES ................................................ iv CHAPTER 1 INTRODUCTION ............................................... 1 1.1 Thesis Statement and Contributions......................2 2 BACKGROUND ................................................ 4 2.1 Statistical Parsing................................4 2.2 Semantic Knowledge Base............................5 2.3 Word Sense Disambiguation...........................6 2.4 COactive Language Definition (COLD)....................6 3 COMPUTING SEMANTIC RELATEDNESS ........................... 8 3.1 Semantic Comparison of Words.........................8 3.2 Dynamic Rule Declarations...........................9 3.3 Comparing Natural Language and Definite Clause Grammar........ 13 4 RESULTS ..................................................... 15 4.1 Semantic Analysis of Natural Language and DCGs.............. 15 4.2 A Framework for Computing Semantic Relatedness.............. 18 5 CONCLUSIONS ................................................ 20 REFERENCES ................................................... 22 iv LIST OF FIGURES Figure Page 2.1 COLD source code sample............................7 3.1 Graph walk in WordNet.............................9 3.2 POS dictionary sample.............................. 10 3.3 Defining almost identical COLD rules..................... 11 3.4 Nested rule declarations............................. 12 3.5 Simple dictionary sample............................ 12 3.6 Using dictionary-defined parameters...................... 12 3.7 Using multiple dictionary-defined parameters in one rule........... 13 4.1 COLD semantic DCG sample.......................... 15 4.2 A dictionary sample............................... 16 4.3 A natural sentence to be used as an input................... 16 4.4 Test application.................................. 19 CHAPTER 1 INTRODUCTION Applications for computing semantic similarity are becoming more prevalent [1, 5]. A semantic similarity is a comparison of how similar in meaning two concepts are. These concepts are associated with some ambiguity, normally related to natural language. Com- parisons could be between: two words, two natural sentences, or between a sentence and a predefined statement (a rule). Such a rule would have some grammatical structure, e.g., its grammar could be described as a set of definite clauses in first-order logic. Such a repre- sentation is denoted Definite Clause Grammar (DCG). A problem with DCGs is that they are static as to what they can express. Every combination of possible words needs to be explicitly defined in order to be acceptable rules. This is a complex problem for systems comparing a user’s natural language with DCGs, because a user cannot be expected to con- struct sentences exactly matching rules. As such, rules are defined with close consideration of natural language, as can be seen in [3]. Conventional approaches for comparing texts fail to deliver human-level (common sense) results [24]. This is understandable due to the many different ways semantically identical sentences can be expressed using natural language. Computed semantic similarity measurements are used in a wide range of services which rely on natural language. There- fore, increasing the confidence of which texts can be compared is desirable in many natural language processing applications, including: machine translation [20], conversational agents [23], web-page retrieval (e.g., by search engines) and image retrieval [21, 19]. This thesis is part of larger project within the field of human robot interaction, aiming to decrease the drawback of robot deployment for small and medium enterprises (SMEs). The investment cost of deploying robots for SMEs is partly related to hardware purchases, but also the required cost of contracting expert robot programmers. A high level natural 2 language framework is purposed through [4] which aims to remove the dependencies on the robot programmers, allowing for such tasks, and (re)purposing, to be performed by manufacturing engineers. COLD, or COactive Language Definition, is a high level programming language for the rapid development of multimodal interfaces [3]. In its current version, possible multimodal sentences are described as context free definite clause grammar. That is, every word that is desired to be acceptable by COLD needs to be explicitly written as a rule in a COLD grammar file. This limits the capabilities of the language and make programming it cum- bersome. For instance, the sentences: ”go to your house” and ”go home” are similar both semantically and lexically. However, two different rules must be defined to be able to handle both cases. 1.1 Thesis Statement and Contributions This thesis contributes to the field of natural language processing. Specifically, it discusses the benefits of extending static definite clause grammar (DCG) systems–that match users’ inputs with predefined rules–with tools for semantic analysis. Also, it supplies a modular framework for DCG–text and text–text comparisons. Although said framework is not the purpose of this thesis, it can serve as a basis for further development and verification. The purpose of this thesis is to: 1. present an approach for computing semantic confidence measurements between natu- ral lingual phrases and DCGs, and 2. show how an existing static DCG-based system can be extended with said function- ality. Our algorithm (Section 4.1) combines common natural language processing tasks such as: statistical parsing, tokenization, parts-of-speech tagging and word-sense disambiguation. We apply these techniques through the context of the combined word-sense of the input and DCGs. We extend the understood meaning of concepts using large lexical and semantic 3 thesauri. Doing so, our algorithm successfully match linguistic phrases with (somewhat) ambiguous predefined grammar. We show how our approach can benefit DCG-bases system. This includes: • Greater freedom in the definitions of the grammar rules. • Not requiring all usable words to be predefined. • Parts of parsing can be shifted to the statistical parser, allowing for early termination of parses where further traversing of the parse tree would not otherwise be beneficial. Furthermore, this could enable rule definitions to be performed without the explicit con- sideration of users’ natural language, but rather in a way that better conveys the semantic goal of the rule. Although our approach focuses on DCG-based systems, it is still applicable in any system where the semantic relatedness of two texts is desired, because in essence, a semantic DCG is a text. 4 CHAPTER 2 BACKGROUND Before a meaningful semantic comparison can be performed on two sentences of natural text, they first need to be parsed. This process involves chunking the sentences and tagging the individual words of a given phrase with its corresponding part-of-speech
Recommended publications
  • Some Strands of Wittgenstein's Normative Pragmatism
    Some Strands of Wittgenstein’s Normative Pragmatism, and Some Strains of his Semantic Nihilism ROBERT B. BRANDOM ABSTRACT WORK TYPE In this reflection I address one of the critical questions this monograph is Article about: How to justify proposing yet another semantic theory in the light of Wittgenstein’s strong warnings against it. I see two clear motives for ARTICLE HISTORY Wittgenstein’s semantic nihilism. The first one is the view that philosophical Received: problems arise from postulating hypothetical entities such as ‘meanings’. To 27–January–2018 dissolve the philosophical problems rather than create new ones, Wittgenstein Accepted: suggests substituting ‘meaning’ with ‘use’ and avoiding scientism in philosophy 3–March–2018 together with the urge to penetrate in one's investigation to unobservable depths. I believe this first motive constitutes only a weak motive for ARTICLE LANGUAGE Wittgenstein’s quietism, because there are substantial differences between English empirical theories in natural sciences and semantic theories in philosophy that KEYWORDS leave Wittgenstein’s assimilation of both open to criticism. But Wittgenstein is Meaning and Use right, on the second motive, that given the dynamic character of linguistic Hypothetical Entities practice, the classical project of semantic theory is a disease that can be Antiscientism removed or ameliorated only by heeding the advice to replace concern with Semantic Nihilism meaning by concern with use. On my view, this does not preclude, however, a Linguistic Dynamism different kind of theoretical approach to meaning that avoids the pitfalls of the Procrustean enterprise Wittgenstein complained about. © Studia Humanitatis – Universidad de Salamanca 2019 R. Brandom (✉) Disputatio. Philosophical Research Bulletin University of Pittsburgh, USA Vol.
    [Show full text]
  • Semantic Analysis of the First Cities from a Deconstruction Perspective Doi: 10.23968/2500-0055-2020-5-3-43-48
    Mojtaba Valibeigi, Faezeh Ashuri— Pages 43–48 SEMANTIC ANALYSIS OF THE FIRST CITIES FROM A DECONSTRUCTION PERSPECTIVE DOI: 10.23968/2500-0055-2020-5-3-43-48 SEMANTIC ANALYSIS OF THE FIRST CITIES FROM A DECONSTRUCTION PERSPECTIVE Mojtaba Valibeigi*, Faezeh Ashuri Buein Zahra Technical University Imam Khomeini Blvd, Buein Zahra, Qazvin, Iran *Corresponding author: [email protected] Abstract Introduction: Deconstruction is looking for any meaning, semantics and concepts and then shows how all of them seem to lead to chaos, and are always on the border of meaning duality. Purpose of the study: The study is aimed to investigate urban identity from a deconstruction perspective. Since two important bases of deconstruction are text and meaning and their relationships, we chose the first cities on a symbolic level as a text and tried to analyze their meanings. Methods: The study used a deductive content analysis in three steps including preparation, organization and final report or conclusion. In the first step, we argued deconstruction philosophy based on Derrida’s views accordingly. Then we determined some common semantic features of the first cities. Finally, we presented some conclusions based on a semantic interpretation of the first cities’ identity.Results: It seems that all cities as texts tend to provoke a special imaginary meaning, while simultaneously promoting and emphasizing the opposite meaning of what they want to show. Keywords Deconstruction, text, meaning, urban semantics, the first cities. Introduction to have a specific meaning (Gualberto and Kress, 2019; Expressions are an act of recognizing or displaying Leone, 2019; Stojiljković and Ristić Trajković, 2018).
    [Show full text]
  • Matrix Decompositions and Latent Semantic Indexing
    Online edition (c)2009 Cambridge UP DRAFT! © April 1, 2009 Cambridge University Press. Feedback welcome. 403 Matrix decompositions and latent 18 semantic indexing On page 123 we introduced the notion of a term-document matrix: an M N matrix C, each of whose rows represents a term and each of whose column× s represents a document in the collection. Even for a collection of modest size, the term-document matrix C is likely to have several tens of thousands of rows and columns. In Section 18.1.1 we first develop a class of operations from linear algebra, known as matrix decomposition. In Section 18.2 we use a special form of matrix decomposition to construct a low-rank approximation to the term-document matrix. In Section 18.3 we examine the application of such low-rank approximations to indexing and retrieving documents, a technique referred to as latent semantic indexing. While latent semantic in- dexing has not been established as a significant force in scoring and ranking for information retrieval, it remains an intriguing approach to clustering in a number of domains including for collections of text documents (Section 16.6, page 372). Understanding its full potential remains an area of active research. Readers who do not require a refresher on linear algebra may skip Sec- tion 18.1, although Example 18.1 is especially recommended as it highlights a property of eigenvalues that we exploit later in the chapter. 18.1 Linear algebra review We briefly review some necessary background in linear algebra. Let C be an M N matrix with real-valued entries; for a term-document matrix, all × RANK entries are in fact non-negative.
    [Show full text]
  • Linguistic Relativity Hyp
    THE LINGUISTIC RELATIVITY HYPOTHESIS by Michele Nathan A Thesis Submitted to the Faculty of the College of Social Science in Partial Fulfillment of the Requirements for the Degree of Master of Arts Florida Atlantic University Boca Raton, Florida December 1973 THE LINGUISTIC RELATIVITY HYPOTHESIS by Michele Nathan This thesis was prepared under the direction of the candidate's thesis advisor, Dr. John D. Early, Department of Anthropology, and has been approved by the members of his supervisory committee. It was submitted to the faculty of the College of Social Science and was accepted in partial fulfillment of the requirements for the degree of Master of Arts. SUPERVISORY COMMITTEE: &~ rl7 IC?13 (date) 1 ii ABSTRACT Author: Michele Nathan Title: The Linguistic Relativity Hypothesis Institution: Florida Atlantic University Degree: Master of Arts Year: 1973 Although interest in the linguistic relativity hypothesis seems to have waned in recent years, this thesis attempts to assess the available evidence supporting it in order to show that further investigation of the hypothesis might be most profitable. Special attention is paid to the fact that anthropology has largely failed to substantiate any claims that correlations between culture and the semantics of language do exist. This has been due to the impressionistic nature of the studies in this area. The use of statistics and hypothesis testing to provide mor.e rigorous methodology is discussed in the hope that employing such paradigms would enable anthropology to contribute some sound evidence regarding t~~ hypothesis. iii TABLE OF CONTENTS Page Introduction • 1 CHAPTER I THE.HISTORY OF THE FORMULATION OF THE HYPOTHESIS.
    [Show full text]
  • Philosophy of Language in the Twentieth Century Jason Stanley Rutgers University
    Philosophy of Language in the Twentieth Century Jason Stanley Rutgers University In the Twentieth Century, Logic and Philosophy of Language are two of the few areas of philosophy in which philosophers made indisputable progress. For example, even now many of the foremost living ethicists present their theories as somewhat more explicit versions of the ideas of Kant, Mill, or Aristotle. In contrast, it would be patently absurd for a contemporary philosopher of language or logician to think of herself as working in the shadow of any figure who died before the Twentieth Century began. Advances in these disciplines make even the most unaccomplished of its practitioners vastly more sophisticated than Kant. There were previous periods in which the problems of language and logic were studied extensively (e.g. the medieval period). But from the perspective of the progress made in the last 120 years, previous work is at most a source of interesting data or occasional insight. All systematic theorizing about content that meets contemporary standards of rigor has been done subsequently. The advances Philosophy of Language has made in the Twentieth Century are of course the result of the remarkable progress made in logic. Few other philosophical disciplines gained as much from the developments in logic as the Philosophy of Language. In the course of presenting the first formal system in the Begriffsscrift , Gottlob Frege developed a formal language. Subsequently, logicians provided rigorous semantics for formal languages, in order to define truth in a model, and thereby characterize logical consequence. Such rigor was required in order to enable logicians to carry out semantic proofs about formal systems in a formal system, thereby providing semantics with the same benefits as increased formalization had provided for other branches of mathematics.
    [Show full text]
  • A History of Semantics Naming
    1/27 A history of semantics [in Nick Riemer (ed.) Routledge Handbook of Semantics] Naming [M]an “makes” his territory by naming the “things” in it. (Chatwin 1988: 301) Human beings name things in their environment. The name helps to distinguish and identify the denotatum (thing named) and is essential to communication with fellow humans about 1 such denotata. In Plato’s Cratylus (Plato 1997) c. 385 BCE, Socrates advances the hypothesis that the earliest name-giver (onomatourgos) selected a name that captures the essence of its denotatum, that is in some way iconic as with onomatopoeic bird names like cuckoo or whippoorwill. On this hypothesis the meaning of a word would be ‘natural’ because directly recognizable from the form of the word. Many of the Ancients sought to demonstrate that names are far more descriptive than the facts allow. For example Socrates in Cratylus 406c derives the name Dionusos (god of Bacchanalia) from didous ton oinon ‘giving wine’. In De lingua latina V: 101 (c. 45 BCE) Varro suggests that because the fox is fleet-footed, volpes ‘fox’ is a blend of volare ‘fly’ and pes ‘foot’ (Varro 1938). Isidore of Seville suggested c. 625 CE that oratio ‘utterance’ derives from oris ratio ‘the mouth’s reason’. None of these is correct and many such ‘etymologies’ are utterly absurd (see Allan 2010). Indeed, the implausibility of such accounts was recognized by Socrates in Cratylus (426b-427b, 434e- 435c), but a clear statement that names are symbols which denote by convention is first found some 25 years after Cratylus in Aristotle’s On Interpretation 16a3, 16a20 (Aristotle 1984).
    [Show full text]
  • 11 Semantics
    "'If" '.u' $. Semantics 115 ~' used to describe a product, with 'good for you', but we would not want to ., mclude this association within the basic conceptual meaning of the expres­ sion. Poets and advertisers are, of course, very interested in using terms in such a way that their associative meanings are evoked, and some linguists do lfivestigate this aspect of language use. However, in this chapter we shall be more interested in characterizing what constitutes the conceptual meaning vi terms. Semantic features , So, how would a semantic approach help us to understand something about 11 Semantics Ihe nature of language? One way it might be helpful would be as a means of accounting for the 'oddness' we experience when we read English sentences such as the following: I once referred to a character in one of my cartoons as a "dork" (a popular j The hamburgerate the man insult when I was growing up), but myeditor called me up and said that "dork" My cat studiedlinguistics couldn't be used because it meant "penis." I couldn't believe it. I ran to my A table was listening to some music New Dictionary ofAmerican Slang and, sure enough, he was right. Allthose years of saying or being called a "dork" and I had never really known what it ~otice that the oddness of these sentences does not derive from their syn­ meant. What a nerd. Gary Larson (1989) tactic structure. According to some basic syntactic rules for forming English sentences (such as those presented in Chapter 10), we have well-structured Semantics is the study of the meaning of words, phrases and sentences.
    [Show full text]
  • The Leibnizian Vision in Frank Ankersmit's Philosophy of History
    INTENSION, SUBSTANCE AND CALCULUS – THE LEIBNIZIAN VISION IN FRANK ANKERSMIT’S PHILOSOPHY OF HISTORY Oulu University Faculty of History Master’s thesis 20.5.2019 Aleksi Oja 1 Due to the abundance of referral to the extensive work of Frank Ankersmit and my general reliance on his published books and selected papers as sources, I ask the reader to use the following list of abbreviations of some of the most featured works: NL – Ankersmit, F. R. 1983: Narrative Logic – a Semantic Analysis of the Historian’s Language. Martinus Nijhoff publishers, The Hague, Netherlands. HT – Ankersmit, F. R. 1994. History and Tropology: The Rise and Fall of Metaphor. University of California Press. London, England. HR – Ankersmit, F. R. 2001: Historical Representation. Stanford University Press, California, USA. PR – Ankersmit, F. R. 2002: Political Representation. Stanford University Press, California, USA. SHE – Ankersmit, F. R. 2005: Sublime Historical Experience. Stanford University Press, California, USA. MTR – Ankersmit F. R. 2012: Meaning, Truth and Reference in Historical Representation. Stanford University Press, California, USA. HSI – Ankersmit, F. R. 2013: History as the Science of the Individual. Journal of the Philosophy of History Vol. 7 (3), 396 – 425. WEM – Ankersmit, F. R. 2017: Where the extremes meet. A presently still an unpublished paper, given as the handout for Frank Ankersmit’s opening speech at the Seminar of philosophy of history in October 2017 in Oulu university. 2 Table of Contents Introduction ..............................................................................................................................
    [Show full text]
  • Ontology and Information Systems
    Ontology and Information Systems 1 Barry Smith Philosophical Ontology Ontology as a branch of philosophy is the science of what is, of the kinds and structures of objects, properties, events, processes and relations in every area of reality. ‘Ontology’ is often used by philosophers as a synonym for ‘metaphysics’ (literally: ‘what comes after the Physics’), a term which was used by early students of Aristotle to refer to what Aristotle himself called ‘first philosophy’.2 The term ‘ontology’ (or ontologia) was itself coined in 1613, independently, by two philosophers, Rudolf Göckel (Goclenius), in his Lexicon philosophicum and Jacob Lorhard (Lorhardus), in his Theatrum philosophicum. The first occurrence in English recorded by the OED appears in Bailey’s dictionary of 1721, which defines ontology as ‘an Account of being in the Abstract’. Methods and Goals of Philosophical Ontology The methods of philosophical ontology are the methods of philosophy in general. They include the development of theories of wider or narrower scope and the testing and refinement of such theories by measuring them up, either against difficult 1 This paper is based upon work supported by the National Science Foundation under Grant No. BCS-9975557 (“Ontology and Geographic Categories”) and by the Alexander von Humboldt Foundation under the auspices of its Wolfgang Paul Program. Thanks go to Thomas Bittner, Olivier Bodenreider, Anita Burgun, Charles Dement, Andrew Frank, Angelika Franzke, Wolfgang Grassl, Pierre Grenon, Nicola Guarino, Patrick Hayes, Kathleen Hornsby, Ingvar Johansson, Fritz Lehmann, Chris Menzel, Kevin Mulligan, Chris Partridge, David W. Smith, William Rapaport, Daniel von Wachter, Chris Welty and Graham White for helpful comments.
    [Show full text]
  • Semantic Analysis on the Use of English Language Slogan
    SEMANTIC ANALYSIS ON THE USE OF ENGLISH LANGUAGE SLOGAN 1Andi Asrifan 2Like Raskova Octaberlina 3Rini Handayani 1,3Universitas Muhammadiyah Sidenreng Rappang 2Universitas Islam Negeri Maulana Malik Ibrahim, Malang, Indonesia Abstract In this study the objectives to be achieved are: (1) To identify what are the types of English slogans. (2) To identify how the semantic meanings of English slogans. The type of research used is qualitative. While the methods and instruments used to collect and analyze and answer all existing problems are qualitative methods and checklist observations. The process of data analysis in this research is carried out in a process starting from data collection and done intensively until the end of the study. Data that researchers get from observations from several sources such as the internet and journals, namely: (1) in general there are 4 types of slogans such as product / advertising slogans, health slogans, environmental slogans and educational slogans. The types of product / advertising slogans are 6 slogans, health slogans 2, environmental slogans 5 and education slogans totaling 8. (2) semantic meanings of 4 types of slogans from internet and journal sources there are 5 namely conceptual meaning, associative meaning, connotative meaning, affective meaning and stylist. Of the 20 slogans education slogans and products that contain more conceptual meaning. The environmental slogans contain more conceptual and affective meanings. Slogans that contain conceptual meaning in the form of a situation context and based on words that have been conceptualized in the dictionary. Key Words: Analysis Semantic, Semantic Meaning, English Slogan. Language is an important component in life so it is not excessive when language is termed as the heart of human life, because language is a very important role in life to support the interaction process.
    [Show full text]
  • Word Senses and Wordnet Lady Bracknell
    Speech and Language Processing. Daniel Jurafsky & James H. Martin. Copyright © 2021. All rights reserved. Draft of September 21, 2021. CHAPTER 18 Word Senses and WordNet Lady Bracknell. Are your parents living? Jack. I have lost both my parents. Lady Bracknell. To lose one parent, Mr. Worthing, may be regarded as a misfortune; to lose both looks like carelessness. Oscar Wilde, The Importance of Being Earnest ambiguous Words are ambiguous: the same word can be used to mean different things. In Chapter 6 we saw that the word “mouse” has (at least) two meanings: (1) a small rodent, or (2) a hand-operated device to control a cursor. The word “bank” can mean: (1) a financial institution or (2) a sloping mound. In the quote above from his play The Importance of Being Earnest, Oscar Wilde plays with two meanings of “lose” (to misplace an object, and to suffer the death of a close person). We say that the words ‘mouse’ or ‘bank’ are polysemous (from Greek ‘having word sense many senses’, poly- ‘many’ + sema, ‘sign, mark’).1 A sense (or word sense) is a discrete representation of one aspect of the meaning of a word. In this chapter WordNet we discuss word senses in more detail and introduce WordNet, a large online the- saurus —a database that represents word senses—with versions in many languages. WordNet also represents relations between senses. For example, there is an IS-A relation between dog and mammal (a dog is a kind of mammal) and a part-whole relation between engine and car (an engine is a part of a car).
    [Show full text]
  • Semantic Change, Semantic Theory and Linguistic Relativity
    Gábor Györi Semantic Change, Semantic Theory and Linguistic Relativity Series A: General & Theoretical Papers ISSN 1435-6473 Essen: LAUD 1998 (2., unveränderte Auflage 2006) Paper No. 444 Universität Duisburg-Essen Gábor Györi Jannus Pannonius University, Pecs (Hungary) Semantic Change, Semantic Theory and Linguistic Relativity Copyright by the author Reproduced by LAUD 1998 (2., unveränderte Auflage 2006) Linguistic Agency Series A University of Duisburg-Essen General and Theoretical FB Geisteswissenschaften Paper No. 444 Universitätsstr. 12 D- 45117 Essen Order LAUD-papers online: http://www.linse.uni-due.de/linse/laud/index.html Or contact: [email protected] Gábor Györi Semantic Change, Semantic Theory and Linguistic Relativity 0. Abstract The most general cause usually given for semantic and lexical change is the arising of new communicative needs in a speech community. However, communication is also a cognitive activity: we communicate our mental states, the contents of our minds, etc. Communicative needs must go hand in hand with cognitive needs, since there is no linguistic communication without the underlying mental representations. Because of this, semantic change can only be accounted for by a semantic theory which takes into account human conceptualization as a product of our understanding of bodily experience, as does cognitive semantics. Semantic change exploits and utilizes the common understanding and interpretation of experience which exists on the basis of conventionalized conceptual structures shared by the members of a speech community. However, while the semantic structure of a language is the product of conceptualization processes, it is also true that the input for these conceptualization processes is at least partly the semantic structure of that language.
    [Show full text]