Unsupervised Learning for Text-To-Speech Synthesis

Total Page:16

File Type:pdf, Size:1020Kb

Unsupervised Learning for Text-To-Speech Synthesis Unsupervised Learning for Text-to-Speech Synthesis Oliver Watts Thesis submitted for the degree of Doctor of Philosophy The University of Edinburgh 2012 Declaration I have composed this thesis. The work in it is my own unless stated. Oliver Watts ii Acknowledgements I would like to thank the following people: • Simon King and Junichi Yamagishi, for all of their supervision, guidance and advice • All those who provided useful feedback after earlier presentations of this work, in particular Steve Renals and Rob Clark • Everybody at the Centre for Speech Technology Research, for creating such a friendly and stimulating place to work • Karl Isaac for helping set up an experiment using Mechanical Turk, and C´assiaValentini-Botinh~aofor helping to recruit local listeners for experi- ments • Adriana Stan at the University of Cluj-Napoca for providing the raw version of the newstext data used in Chapter 9 (and indeed for having made the whole of her Romanian speech corpus generally available), and for recruiting participants for the evaluation of Romanian systems • Martti Vainio and Antti Suni at the University of Helsinki for making the Finnish database available for the experiments described in Chapter 8 • Antti Suni (again) and Heini Kallio at the University of Helsinki and Tuomo Raitio at Aalto University for recruiting participants for the evaluation of Finnish systems. This work was made possible by a Doctoral Training Award from the Engineering and Physical Sciences Research Council (EPSRC: http://www.epsrc.ac.uk), and has made use of the resources provided by the Edinburgh Compute and Data Facility (ECDF: http://www.ecdf.ed.ac.uk). The ECDF is partially supported by the eDIKT initiative (http://www.edikt.org.uk). iii Abstract This thesis introduces a general method for incorporating the distributional analy- sis of textual and linguistic objects into text-to-speech (TTS) conversion systems. Conventional TTS conversion uses intermediate layers of representation to bridge the gap between text and speech. Collecting the annotated data needed to pro- duce these intermediate layers is a far from trivial task, possibly prohibitively so for languages in which no such resources are in existence. Distributional analysis, in contrast, proceeds in an unsupervised manner, and so enables the creation of systems using textual data that are not annotated. The method therefore aids the building of systems for languages in which conventional linguistic resources are scarce, but is not restricted to these languages. The distributional analysis proposed here places the textual objects analysed in a continuous-valued space, rather than specifying a hard categorisation of those objects. This space is then partitioned during the training of acoustic models for synthesis, so that the models generalise over objects' surface forms in a way that is acoustically relevant. The method is applied to three levels of textual analysis: to the characteri- sation of sub-syllabic units, word units and utterances. Entire systems for three languages (English, Finnish and Romanian) are built with no reliance on manu- ally labelled data or language-specific expertise. Results of a subjective evaluation are presented. iv Contents 1 Introduction 1 1.1 Goals and Approach . .1 1.2 Organisation of this thesis . .2 2 Background 5 2.1 Conventional Systems . .5 2.1.1 Synthesiser Front-end: Linguistic modelling . .6 2.1.2 Synthesiser Back-end: Acoustic Modelling . 13 2.1.3 Training Recipes Used . 25 2.2 Types of Alternative Approach . 27 2.2.1 Conventional Approach . 29 2.2.2 Naive Approach . 30 2.2.3 Speech-Driven Approach . 30 2.2.4 Text-Driven Approach . 31 2.2.5 Approach Taken in this Thesis . 31 2.3 Naivety, language independence, and language typology . 32 3 Benchmark Systems 35 3.1 Introduction . 35 3.2 Experiment 1: Contribution of High-Level Features . 37 3.2.1 Overview . 37 3.2.2 Data Used . 38 3.2.3 Annotation Used . 39 3.2.4 Subjective Evaluation . 41 3.2.5 Systems' Use of Linguistic Features . 42 3.3 Experiment 2: Contribution of Phonemic Transcription . 44 3.3.1 Letter-Based Speech Synthesis . 44 3.3.2 Phonemes: Modelling vs. Generalisation . 47 v vi CONTENTS 3.3.3 Induced Compound Features . 47 3.3.4 Evaluation . 50 3.3.5 Results . 52 3.4 Experiment 3: Contribution of Phonetic Categories . 53 3.4.1 Overview . 53 3.4.2 Systems Trained . 53 3.4.3 Evaluation . 56 3.5 Conclusions . 58 4 Vector Space Models 61 4.1 Vector Space Modelling: an Example . 61 4.2 Vector Space Modelling Applied to TTS . 67 4.2.1 Serial Tree Building: Shortcomings . 67 4.2.2 Distributional{Acoustic Modelling: Vector Space Models and Decision Trees . 69 4.2.3 Strengths of the proposed approach . 70 4.2.4 The Vector Space Models Used in this Thesis . 72 4.2.5 Design Choices . 74 4.3 Previous Work on Vector Space Models of Language . 78 4.3.1 Salton's Vector Space Model . 78 4.3.2 Dimension-Reduced Vector Space Model: Latent Semantic Indexing . 81 4.3.3 Reduced Vector Space Model: Other Projections . 86 4.3.4 Probablistic topic models . 87 4.4 Other Techniques for the Induction of Linguistic Representations . 87 4.4.1 Hierarchical Clustering Approaches . 88 4.4.2 HMM-based Approaches . 90 4.4.3 Graph Partitioning Approaches . 91 4.4.4 Connectionist Approaches . 91 5 Letter Space 95 5.1 Introduction . 95 5.2 A Vector Space Model of Letter Types . 96 5.3 A Vector Space Model of Phoneme Types . 97 5.3.1 Experiment . 99 5.4 Conclusions . 102 CONTENTS vii 6 Word Space 103 6.1 Introduction . 103 6.2 A Vector Space Model for Word Types . 104 6.3 Experiment 1: Word VSM for Phrase-Break Prediction . 108 6.3.1 Background: Phrase-break Prediction . 108 6.3.2 Experiment . 109 6.3.3 Results . 114 6.4 Experiment 2: Word VSM for State-Tying . 117 6.4.1 Experiment . 117 6.5 Conclusions and Open Problems . 120 7 Refinements and Extensions 123 7.1 Word Space: Corpus Size and Parameter Settings . 123 7.1.1 Data and Vector Space Models . 124 7.1.2 Results . 126 7.2 Feature Selection for HMM TTS . 129 7.2.1 Ensembles of Trees . 129 7.2.2 Variable Importance Measures . 130 7.2.3 Feature Selection for the Distributional{Acoustic Method . 131 7.2.4 Experiment: Word VSM for State-Tying, Revisited . 135 7.2.5 Conclusion . 135 7.3 An Utterance Space Model . 136 7.3.1 A Vector Space Model of Utterances . 137 7.3.2 Experiment: Separability of Questions and Non-Questions 138 8 Evaluation of Entire Systems 143 8.1 Introduction . 143 8.2 Language Characteristics . 144 8.3 Initial Hypotheses and Systems Built . 146 8.4 System Construction and Features Used . 149 8.4.1 Feature-set A . 149 8.4.2 Feature-sets B, C & D . 154 8.4.3 Feature-set T: Topline Features . 157 8.4.4 Feature Selection: Systems *-*-2 . 157 8.5 Synthesis . 158 8.6 Objective Evaluation . 158 viii CONTENTS 8.6.1 Method . 158 8.6.2 Results . 159 8.7 Subjective Evaluation . 163 8.7.1 Selected Hypotheses . 163 8.7.2 Procedure . 164 8.7.3 Results . 165 8.8 Conclusions . 167 9 Conclusions and Future Work 169 9.1 Contributions . 169 9.2 Future work . 171 List of Figures 2.1 Splitting a toy dataset for letter-to-sound conversion . 11 3.1 Results of AB test for naturalness in Experiment 1 . 42 3.2 Questions asked at test time in Experiment 1 . 43 3.3 Serial tree building: a toy example . 48 3.4 Results of Experiment 2 . 51 3.5 Serial tree building: another toy example . 54 3.6 Results of Experiment 3 . 56 3.7 Leaf node sizes in three systems . 57 4.1 Induction of word representations: toy example . 65 4.2 Two dimensions of a space of possible vector space models . 76 5.1 Two dimensions of a letter space . 96 5.2 Two dimensions of a phoneme space . 98 5.3 Results of phoneme-space experiment . 101 6.1 Results of phrase-break experiment . 114 6.2 Sizes of trees built for phrase-break experiment . 115 6.3 Portions of two trees built for phrase-break experiment . 116 6.4 Results of state-tying experiment . 119 7.1 The effect of varying method for handling unseen words in a word space . 126 7.2 The effect of varying context vocabulary size in a word space . 127 7.3 Results on phrase-break task, compared with earlier results . 128 7.4 Results of state-tying experiment . 134 7.5 Two dimensions of an utterance space (1) . 139 7.6 Two dimensions of an utterance space (2) . 140 8.1 Morphological richness of three target languages . 145 ix x LIST OF FIGURES 8.2 Construction of naive systems . 150 8.3 Objective evaluation results for 15 English, Finnish and Romanian voices . 161 (a) Bark cepstral distortion (English) ............... 161 (b) Root Mean Square Error of F0 (English) ........... 161 (c) Bark cepstral distortion (Finnish) ............... 161 (d) Root Mean Square Error of F0 (Finnish) ........... 161 (e) Bark cepstral distortion (Romanian) ............. 161 (f) Root Mean Square Error of F0 (Romanian) ......... 161 8.4 Objective evaluation results for 30 English, Finnish and Romanian voices . 162 (a) Bark cepstral distortion (English) ............... 162 (b) Root Mean Square Error of F0 (English) ........... 162 (c) Bark cepstral distortion (Finnish) ............... 162 (d) Root Mean Square Error of F0 (Finnish) ........... 162 (e) Bark cepstral distortion (Romanian) ..............
Recommended publications
  • Combining Semantic and Lexical Measures to Evaluate Medical Terms Similarity?
    Combining semantic and lexical measures to evaluate medical terms similarity? Silvio Domingos Cardoso1;2, Marcos Da Silveira1, Ying-Chi Lin3, Victor Christen3, Erhard Rahm3, Chantal Reynaud-Dela^ıtre2, and C´edricPruski1 1 LIST, Luxembourg Institute of Science and Technology, Luxembourg fsilvio.cardoso,marcos.dasilveira,[email protected] 2 LRI, University of Paris-Sud XI, France [email protected] 3 Department of Computer Science, Universit¨atLeipzig, Germany flin,christen,[email protected] Abstract. The use of similarity measures in various domains is corner- stone for different tasks ranging from ontology alignment to information retrieval. To this end, existing metrics can be classified into several cate- gories among which lexical and semantic families of similarity measures predominate but have rarely been combined to complete the aforemen- tioned tasks. In this paper, we propose an original approach combining lexical and ontology-based semantic similarity measures to improve the evaluation of terms relatedness. We validate our approach through a set of experiments based on a corpus of reference constructed by domain experts of the medical field and further evaluate the impact of ontology evolution on the used semantic similarity measures. Keywords: Similarity measures · Ontology evolution · Semantic Web · Medical terminologies 1 Introduction Measuring the similarity between terms is at the heart of many research inves- tigations. In ontology matching, similarity measures are used to evaluate the relatedness between concepts from different ontologies [9]. The outcomes are the mappings between the ontologies, increasing the coverage of domain knowledge and optimize semantic interoperability between information systems. In informa- tion retrieval, similarity measures are used to evaluate the relatedness between units of language (e.g., words, sentences, documents) to optimize search [39].
    [Show full text]
  • The Role of Higher-Level Linguistic Features in HMM-Based Speech Synthesis
    INTERSPEECH 2010 The role of higher-level linguistic features in HMM-based speech synthesis Oliver Watts, Junichi Yamagishi, Simon King Centre for Speech Technology Research, University of Edinburgh, UK [email protected] [email protected] [email protected] Abstract the annotation of a synthesiser’s training data on listeners’ re- action to the speech produced by that synthesiser is still unclear We analyse the contribution of higher-level elements of the lin- to us. In particular, we suspect that features such as pitch ac- guistic specification of a data-driven speech synthesiser to the cent type which might be useful in voice-building if labelled naturalness of the synthetic speech which it generates. The reliably, are under-used in conventional systems because of la- system is trained using various subsets of the full feature-set, belling/prediction errors. For building the systems to be evalu- in which features relating to syntactic category, intonational ated we therefore used data for which hand labelled annotation phrase boundary, pitch accent and boundary tones are selec- of ToBI events is available. This allows us to compare ideal sys- tively removed. Utterances synthesised by the different config- tems built from a corpus where these higher-level features are urations of the system are then compared in a subjective evalu- accurately annotated with more conventional systems that rely ation of their naturalness. exclusively on prediction of these features from text for annota- The work presented forms background analysis for an on- tion of their training data. going set of experiments in performing text-to-speech (TTS) conversion based on shallow features: features that can be triv- 2.
    [Show full text]
  • Evaluating Lexical Similarity to Build Sentiment Similarity
    Evaluating Lexical Similarity to Build Sentiment Similarity Grégoire Jadi∗, Vincent Claveauy, Béatrice Daille∗, Laura Monceaux-Cachard∗ ∗ LINA - Univ. Nantes {gregoire.jadi beatrice.daille laura.monceaux}@univ-nantes.fr y IRISA–CNRS, France [email protected] Abstract In this article, we propose to evaluate the lexical similarity information provided by word representations against several opinion resources using traditional Information Retrieval tools. Word representation have been used to build and to extend opinion resources such as lexicon, and ontology and their performance have been evaluated on sentiment analysis tasks. We question this method by measuring the correlation between the sentiment proximity provided by opinion resources and the semantic similarity provided by word representations using different correlation coefficients. We also compare the neighbors found in word representations and list of similar opinion words. Our results show that the proximity of words in state-of-the-art word representations is not very effective to build sentiment similarity. Keywords: opinion lexicon evaluation, sentiment analysis, spectral representation, word embedding 1. Introduction tend or build sentiment lexicons. For example, (Toh and The goal of opinion mining or sentiment analysis is to iden- Wang, 2014) uses the syntactic categories from WordNet tify and classify texts according to the opinion, sentiment or as features for a CRF to extract aspects and WordNet re- emotion that they convey. It requires a good knowledge of lations such as antonymy and synonymy to extend an ex- the sentiment properties of each word. This properties can isting opinion lexicons. Similarly, (Agarwal et al., 2011) be either discovered automatically by machine learning al- look for synonyms in WordNet to find the polarity of words gorithms, or embedded into language resources.
    [Show full text]
  • Using Prior Knowledge to Guide BERT's Attention in Semantic
    Using Prior Knowledge to Guide BERT’s Attention in Semantic Textual Matching Tasks Tingyu Xia Yue Wang School of Artificial Intelligence, Jilin University School of Information and Library Science Key Laboratory of Symbolic Computation and Knowledge University of North Carolina at Chapel Hill Engineering, Jilin University [email protected] [email protected] Yuan Tian∗ Yi Chang∗ School of Artificial Intelligence, Jilin University School of Artificial Intelligence, Jilin University Key Laboratory of Symbolic Computation and Knowledge International Center of Future Science, Jilin University Engineering, Jilin University Key Laboratory of Symbolic Computation and Knowledge [email protected] Engineering, Jilin University [email protected] ABSTRACT 1 INTRODUCTION We study the problem of incorporating prior knowledge into a deep Measuring semantic similarity between two pieces of text is a fun- Transformer-based model, i.e., Bidirectional Encoder Representa- damental and important task in natural language understanding. tions from Transformers (BERT), to enhance its performance on Early works on this task often leverage knowledge resources such semantic textual matching tasks. By probing and analyzing what as WordNet [28] and UMLS [2] as these resources contain well- BERT has already known when solving this task, we obtain better defined types and relations between words and concepts. Recent understanding of what task-specific knowledge BERT needs the works have shown that deep learning models are more effective most and where it is most needed. The analysis further motivates on this task by learning linguistic knowledge from large-scale text us to take a different approach than most existing works. Instead of data and representing text (words and sentences) as continuous using prior knowledge to create a new training task for fine-tuning trainable embedding vectors [10, 17].
    [Show full text]
  • Latent Semantic Analysis for Text-Based Research
    Behavior Research Methods, Instruments, & Computers 1996, 28 (2), 197-202 ANALYSIS OF SEMANTIC AND CLINICAL DATA Chaired by Matthew S. McGlone, Lafayette College Latent semantic analysis for text-based research PETER W. FOLTZ New Mexico State University, Las Cruces, New Mexico Latent semantic analysis (LSA) is a statistical model of word usage that permits comparisons of se­ mantic similarity between pieces of textual information. This papersummarizes three experiments that illustrate how LSA may be used in text-based research. Two experiments describe methods for ana­ lyzinga subject's essay for determining from what text a subject learned the information and for grad­ ing the quality of information cited in the essay. The third experiment describes using LSAto measure the coherence and comprehensibility of texts. One of the primary goals in text-comprehension re­ A theoretical approach to studying text comprehen­ search is to understand what factors influence a reader's sion has been to develop cognitive models ofthe reader's ability to extract and retain information from textual ma­ representation ofthe text (e.g., Kintsch, 1988; van Dijk terial. The typical approach in text-comprehension re­ & Kintsch, 1983). In such a model, semantic information search is to have subjects read textual material and then from both the text and the reader's summary are repre­ have them produce some form of summary, such as an­ sented as sets ofsemantic components called propositions. swering questions or writing an essay. This summary per­ Typically, each clause in a text is represented by a single mits the experimenter to determine what information the proposition.
    [Show full text]
  • The RACAI Text-To-Speech Synthesis System
    The RACAI Text-to-Speech Synthesis System Tiberiu Boroș, Radu Ion, Ștefan Daniel Dumitrescu Research Institute for Artificial Intelligence “Mihai Drăgănescu”, Romanian Academy (RACAI) [email protected], [email protected], [email protected] of standalone Natural Language Processing (NLP) Tools Abstract aimed at enabling text-to-speech (TTS) synthesis for less- This paper describes the RACAI Text-to-Speech (TTS) entry resourced languages. A good example is the case of for the Blizzard Challenge 2013. The development of the Romanian, a language which poses a lot of challenges for TTS RACAI TTS started during the Metanet4U project and the synthesis mainly because of its rich morphology and its system is currently part of the METASHARE platform. This reduced support in terms of freely available resources. Initially paper describes the work carried out for preparing the RACAI all the tools were standalone, but their design allowed their entry during the Blizzard Challenge 2013 and provides a integration into a single package and by adding a unit selection detailed description of our system and future development speech synthesis module based on the Pitch Synchronous directions. Overlap-Add (PSOLA) algorithm we were able to create a fully independent text-to-speech synthesis system that we refer Index Terms: speech synthesis, unit selection, concatenative to as RACAI TTS. 1. Introduction A considerable progress has been made since the start of the project, but our system still requires development and Text-to-speech (TTS) synthesis is a complex process that although considered premature, our participation in the addresses the task of converting arbitrary text into voice.
    [Show full text]
  • Ying Liu, Xi Yang
    Liu, Y. & Yang, X. (2018). A similar legal case retrieval Liu & Yang system by multiple speech question and answer. In Proceedings of The 18th International Conference on Electronic Business (pp. 72-81). ICEB, Guilin, China, December 2-6. A Similar Legal Case Retrieval System by Multiple Speech Question and Answer (Full paper) Ying Liu, School of Computer Science and Information Engineering, Guangxi Normal University, China, [email protected] Xi Yang*, School of Computer Science and Information Engineering, Guangxi Normal University, China, [email protected] ABSTRACT In real life, on the one hand people often lack legal knowledge and legal awareness; on the other hand lawyers are busy, time- consuming, and expensive. As a result, a series of legal consultation problems cannot be properly handled. Although some legal robots can answer some problems about basic legal knowledge that are matched by keywords, they cannot do similar case retrieval and sentencing prediction according to the criminal facts described in natural language. To overcome the difficulty, we propose a similar case retrieval system based on natural language understanding. The system uses online speech synthesis of IFLYTEK and speech reading and writing technology, integrates natural language semantic processing technology and multiple rounds of question-and-answer dialogue mechanism to realise the legal knowledge question and answer with the memory-based context processing ability, and finally retrieves a case that is most similar to the criminal facts that the user consulted. After trial use, the system has a good level of human-computer interaction and high accuracy of information retrieval, which can largely satisfy people's consulting needs for legal issues.
    [Show full text]
  • Untangling Semantic Similarity: Modeling Lexical Processing Experiments with Distributional Semantic Models
    Untangling Semantic Similarity: Modeling Lexical Processing Experiments with Distributional Semantic Models. Farhan Samir Barend Beekhuizen Suzanne Stevenson Department of Computer Science Department of Language Studies Department of Computer Science University of Toronto University of Toronto, Mississauga University of Toronto ([email protected]) ([email protected]) ([email protected]) Abstract DSMs thought to instantiate one but not the other kind of sim- Distributional semantic models (DSMs) are substantially var- ilarity have been found to explain different priming effects on ied in the types of semantic similarity that they output. Despite an aggregate level (as opposed to an item level). this high variance, the different types of similarity are often The underlying conception of semantic versus associative conflated as a monolithic concept in models of behavioural data. We apply the insight that word2vec’s representations similarity, however, has been disputed (Ettinger & Linzen, can be used for capturing both paradigmatic similarity (sub- 2016; McRae et al., 2012), as has one of the main ways in stitutability) and syntagmatic similarity (co-occurrence) to two which it is operationalized (Gunther¨ et al., 2016). In this pa- sets of experimental findings (semantic priming and the effect of semantic neighbourhood density) that have previously been per, we instead follow another distinction, based on distri- modeled with monolithic conceptions of DSM-based seman- butional properties (e.g. Schutze¨ & Pedersen, 1993), namely, tic similarity. Using paradigmatic and syntagmatic similarity that between syntagmatically related words (words that oc- based on word2vec, we show that for some tasks and types of items the two types of similarity play complementary ex- cur in each other’s near proximity, such as drink–coffee), planatory roles, whereas for others, only syntagmatic similar- and paradigmatically related words (words that can be sub- ity seems to matter.
    [Show full text]
  • Synthesis and Recognition of Speech Creating and Listening to Speech
    ISSN 1883-1974 (Print) ISSN 1884-0787 (Online) National Institute of Informatics News NII Interview 51 A Combination of Speech Synthesis and Speech Oct. 2014 Recognition Creates an Affluent Society NII Special 1 “Statistical Speech Synthesis” Technology with a Rapidly Growing Application Area NII Special 2 Finding Practical Application for Speech Recognition Feature Synthesis and Recognition of Speech Creating and Listening to Speech A digital book version of “NII Today” is now available. http://www.nii.ac.jp/about/publication/today/ This English language edition NII Today corresponds to No. 65 of the Japanese edition [Advance Notice] Great news! NII Interview Yamagishi-sensei will create my voice! Yamagishi One result is a speech translation sys- sound. Bit (NII Character) A Combination of Speech Synthesis tem. This system recognizes speech and translates it Ohkawara I would like your comments on the fu- using machine translation to synthesize speech, also ture challenges. and Speech Recognition automatically translating it into every language to Ono The challenge for speech recognition is how speak. Moreover, the speech is created with a human close it will come to humans in the distant speech A Word from the Interviewer voice. In second language learning, you can under- case. If this study is advanced, it will be possible to Creates an Affluent Society stand how you should pronounce it with your own summarize the contents of a meeting and to automati- voice. If this is further advanced, the system could cally take the minutes. If a robot understands the con- have an actor in a movie speak in a different language tents of conversations by multiple people in a natural More and more people have begun to use smart- a smartphone, I use speech input more often.
    [Show full text]
  • The Role of Speech Processing in Human-Computer Intelligent Communication
    THE ROLE OF SPEECH PROCESSING IN HUMAN-COMPUTER INTELLIGENT COMMUNICATION Candace Kamm, Marilyn Walker and Lawrence Rabiner Speech and Image Processing Services Research Laboratory AT&T Labs-Research, Florham Park, NJ 07932 Abstract: We are currently in the midst of a revolution in communications that promises to provide ubiquitous access to multimedia communication services. In order to succeed, this revolution demands seamless, easy-to-use, high quality interfaces to support broadband communication between people and machines. In this paper we argue that spoken language interfaces (SLIs) are essential to making this vision a reality. We discuss potential applications of SLIs, the technologies underlying them, the principles we have developed for designing them, and key areas for future research in both spoken language processing and human computer interfaces. 1. Introduction Around the turn of the twentieth century, it became clear to key people in the Bell System that the concept of Universal Service was rapidly becoming technologically feasible, i.e., the dream of automatically connecting any telephone user to any other telephone user, without the need for operator assistance, became the vision for the future of telecommunications. Of course a number of very hard technical problems had to be solved before the vision could become reality, but by the end of 1915 the first automatic transcontinental telephone call was successfully completed, and within a very few years the dream of Universal Service became a reality in the United States. We are now in the midst of another revolution in communications, one which holds the promise of providing ubiquitous service in multimedia communications.
    [Show full text]
  • Measuring Semantic Similarity of Words Using Concept Networks
    Measuring semantic similarity of words using concept networks Gabor´ Recski Eszter Iklodi´ Research Institute for Linguistics Dept of Automation and Applied Informatics Hungarian Academy of Sciences Budapest U of Technology and Economics H-1068 Budapest, Benczur´ u. 33 H-1117 Budapest, Magyar tudosok´ krt. 2 [email protected] [email protected] Katalin Pajkossy Andras´ Kornai Department of Algebra Institute for Computer Science Budapest U of Technology and Economics Hungarian Academy of Sciences H-1111 Budapest, Egry J. u. 1 H-1111 Budapest, Kende u. 13-17 [email protected] [email protected] Abstract using word embeddings and WordNet. In Sec- tion 3 we briefly introduce the 4lang resources We present a state-of-the-art algorithm and the formalism it uses for encoding the mean- for measuring the semantic similarity of ing of words as directed graphs of concepts, then word pairs using novel combinations of document our efforts to develop novel 4lang- word embeddings, WordNet, and the con- based similarity features. Besides improving the cept dictionary 4lang. We evaluate our performance of existing systems for measuring system on the SimLex-999 benchmark word similarity, the goal of the present project is to data. Our top score of 0.76 is higher than examine the potential of 4lang representations in any published system that we are aware of, representing non-trivial lexical relationships that well beyond the average inter-annotator are beyond the scope of word embeddings and agreement of 0.67, and close to the 0.78 standard linguistic ontologies. average correlation between a human rater Section 4 presents our results and pro- and the average of all other ratings, sug- vides rough error analysis.
    [Show full text]
  • Semantic Similarity Between Sentences
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395 -0056 Volume: 04 Issue: 01 | Jan -2017 www.irjet.net p-ISSN: 2395-0072 SEMANTIC SIMILARITY BETWEEN SENTENCES PANTULKAR SRAVANTHI1, DR. B. SRINIVASU2 1M.tech Scholar Dept. of Computer Science and Engineering, Stanley College of Engineering and Technology for Women, Telangana- Hyderabad, India 2Associate Professor - Dept. of Computer Science and Engineering, Stanley College of Engineering and Technology for Women, Telangana- Hyderabad, India ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract - The task of measuring sentence similarity is [1].Sentence similarity is one of the core elements of Natural defined as determining how similar the meanings of two Language Processing (NLP) tasks such as Recognizing sentences are. Computing sentence similarity is not a trivial Textual Entailment (RTE)[2] and Paraphrase Recognition[3]. task, due to the variability of natural language - expressions. Given two sentences, the task of measuring sentence Measuring semantic similarity of sentences is closely related to similarity is defined as determining how similar the meaning semantic similarity between words. It makes a relationship of two sentences is. The higher the score, the more similar between a word and the sentence through their meanings. The the meaning of the two sentences. WordNet and similarity intention is to enhance the concepts of semantics over the measures play an important role in sentence level similarity syntactic measures that are able to categorize the pair of than document level[4]. sentences effectively. Semantic similarity plays a vital role in Natural language processing, Informational Retrieval, Text 1.1 Problem Description Mining, Q & A systems, text-related research and application area.
    [Show full text]