Evaluating Vector-Space Models of Word Representation, Or, the Unreasonable Effectiveness of Counting Words Near Other Words
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Soft Similarity and Soft Cosine Measure: Similarity of Features in Vector Space Model
Soft Similarity and Soft Cosine Measure: Similarity of Features in Vector Space Model Grigori Sidorov1, Alexander Gelbukh1, Helena Gomez-Adorno1, and David Pinto2 1 Centro de Investigacion en Computacion, Instituto Politecnico Nacional, Mexico D.F., Mexico 2 Facultad de Ciencias de la Computacion, Benemerita Universidad Autonoma de Puebla, Puebla, Mexico {sidorov,gelbukh}@cic.ipn.mx, [email protected], [email protected] Abstract. We show how to consider similarity between 1 Introduction features for calculation of similarity of objects in the Vec tor Space Model (VSM) for machine learning algorithms Computation of similarity of specific objects is a basic and other classes of methods that involve similarity be task of many methods applied in various problems in tween objects. Unlike LSA, we assume that similarity natural language processing and many other fields. In between features is known (say, from a synonym dictio natural language processing, text similarity plays crucial nary) and does not need to be learned from the data. role in many tasks from plagiarism detection [18] and We call the proposed similarity measure soft similarity. question answering [3] to sentiment analysis [14-16]. Similarity between features is common, for example, in The most common manner to represent objects is natural language processing: words, n-grams, or syn the Vector Space Model (VSM) [17]. In this model, the tactic n-grams can be somewhat different (which makes objects are represented as vectors of values of features. them different features) but still have much in common: The features characterize each object and have numeric for example, words “play” and “game” are different but values. -
Document and Topic Models: Plsa
10/4/2018 Document and Topic Models: pLSA and LDA Andrew Levandoski and Jonathan Lobo CS 3750 Advanced Topics in Machine Learning 2 October 2018 Outline • Topic Models • pLSA • LSA • Model • Fitting via EM • pHITS: link analysis • LDA • Dirichlet distribution • Generative process • Model • Geometric Interpretation • Inference 2 1 10/4/2018 Topic Models: Visual Representation Topic proportions and Topics Documents assignments 3 Topic Models: Importance • For a given corpus, we learn two things: 1. Topic: from full vocabulary set, we learn important subsets 2. Topic proportion: we learn what each document is about • This can be viewed as a form of dimensionality reduction • From large vocabulary set, extract basis vectors (topics) • Represent document in topic space (topic proportions) 푁 퐾 • Dimensionality is reduced from 푤푖 ∈ ℤ푉 to 휃 ∈ ℝ • Topic proportion is useful for several applications including document classification, discovery of semantic structures, sentiment analysis, object localization in images, etc. 4 2 10/4/2018 Topic Models: Terminology • Document Model • Word: element in a vocabulary set • Document: collection of words • Corpus: collection of documents • Topic Model • Topic: collection of words (subset of vocabulary) • Document is represented by (latent) mixture of topics • 푝 푤 푑 = 푝 푤 푧 푝(푧|푑) (푧 : topic) • Note: document is a collection of words (not a sequence) • ‘Bag of words’ assumption • In probability, we call this the exchangeability assumption • 푝 푤1, … , 푤푁 = 푝(푤휎 1 , … , 푤휎 푁 ) (휎: permutation) 5 Topic Models: Terminology (cont’d) • Represent each document as a vector space • A word is an item from a vocabulary indexed by {1, … , 푉}. We represent words using unit‐basis vectors. -
Redalyc.Latent Dirichlet Allocation Complement in the Vector Space Model for Multi-Label Text Classification
International Journal of Combinatorial Optimization Problems and Informatics E-ISSN: 2007-1558 [email protected] International Journal of Combinatorial Optimization Problems and Informatics México Carrera-Trejo, Víctor; Sidorov, Grigori; Miranda-Jiménez, Sabino; Moreno Ibarra, Marco; Cadena Martínez, Rodrigo Latent Dirichlet Allocation complement in the vector space model for Multi-Label Text Classification International Journal of Combinatorial Optimization Problems and Informatics, vol. 6, núm. 1, enero-abril, 2015, pp. 7-19 International Journal of Combinatorial Optimization Problems and Informatics Morelos, México Available in: http://www.redalyc.org/articulo.oa?id=265239212002 How to cite Complete issue Scientific Information System More information about this article Network of Scientific Journals from Latin America, the Caribbean, Spain and Portugal Journal's homepage in redalyc.org Non-profit academic project, developed under the open access initiative © International Journal of Combinatorial Optimization Problems and Informatics, Vol. 6, No. 1, Jan-April 2015, pp. 7-19. ISSN: 2007-1558. Latent Dirichlet Allocation complement in the vector space model for Multi-Label Text Classification Víctor Carrera-Trejo1, Grigori Sidorov1, Sabino Miranda-Jiménez2, Marco Moreno Ibarra1 and Rodrigo Cadena Martínez3 Centro de Investigación en Computación1, Instituto Politécnico Nacional, México DF, México Centro de Investigación e Innovación en Tecnologías de la Información y Comunicación (INFOTEC)2, Ags., México Universidad Tecnológica de México3 – UNITEC MÉXICO [email protected], {sidorov,marcomoreno}@cic.ipn.mx, [email protected] [email protected] Abstract. In text classification task one of the main problems is to choose which features give the best results. Various features can be used like words, n-grams, syntactic n-grams of various types (POS tags, dependency relations, mixed, etc.), or a combinations of these features can be considered. -
Gensim Is Robust in Nature and Has Been in Use in Various Systems by Various People As Well As Organisations for Over 4 Years
Gensim i Gensim About the Tutorial Gensim = “Generate Similar” is a popular open source natural language processing library used for unsupervised topic modeling. It uses top academic models and modern statistical machine learning to perform various complex tasks such as Building document or word vectors, Corpora, performing topic identification, performing document comparison (retrieving semantically similar documents), analysing plain-text documents for semantic structure. Audience This tutorial will be useful for graduates, post-graduates, and research students who either have an interest in Natural Language Processing (NLP), Topic Modeling or have these subjects as a part of their curriculum. The reader can be a beginner or an advanced learner. Prerequisites The reader must have basic knowledge about NLP and should also be aware of Python programming concepts. Copyright & Disclaimer Copyright 2020 by Tutorials Point (I) Pvt. Ltd. All the content and graphics published in this e-book are the property of Tutorials Point (I) Pvt. Ltd. The user of this e-book is prohibited to reuse, retain, copy, distribute or republish any contents or a part of contents of this e-book in any manner without written consent of the publisher. We strive to update the contents of our website and tutorials as timely and as precisely as possible, however, the contents may contain inaccuracies or errors. Tutorials Point (I) Pvt. Ltd. provides no guarantee regarding the accuracy, timeliness or completeness of our website or its contents including this tutorial. If you discover any errors on our website or in this tutorial, please notify us at [email protected] ii Gensim Table of Contents About the Tutorial .......................................................................................................................................... -
Deconstructing Word Embedding Models
Deconstructing Word Embedding Models Koushik K. Varma Ashoka University, [email protected] Abstract – Analysis of Word Embedding Models through Analysis, one must be certain that all linguistic aspects of a a deconstructive approach reveals their several word are captured within the vector representations because shortcomings and inconsistencies. These include any discrepancies could be further amplified in practical instability of the vector representations, a distorted applications. While these vector representations have analogical reasoning, geometric incompatibility with widespread use in modern natural language processing, it is linguistic features, and the inconsistencies in the corpus unclear as to what degree they accurately encode the data. A new theoretical embedding model, ‘Derridian essence of language in its structural, contextual, cultural and Embedding,’ is proposed in this paper. Contemporary ethical assimilation. Surface evaluations on training-test embedding models are evaluated qualitatively in terms data provide efficiency measures for a specific of how adequate they are in relation to the capabilities of implementation but do not explicitly give insights on the a Derridian Embedding. factors causing the existing efficiency (or the lack there of). We henceforth present a deconstructive review of 1. INTRODUCTION contemporary word embeddings using available literature to clarify certain misconceptions and inconsistencies Saussure, in the Course in General Linguistics, concerning them. argued that there is no substance in language and that all language consists of differences. In language there are only forms, not substances, and by that he meant that all 1.1 DECONSTRUCTIVE APPROACH apparently substantive units of language are generated by other things that lie outside them, but these external Derrida, a post-structualist French philosopher, characteristics are actually internal to their make-up. -
Vector Space Models for Words and Documents Statistical Natural Language Processing ELEC-E5550 Jan 30, 2019 Tiina Lindh-Knuutila D.Sc
Vector space models for words and documents Statistical Natural Language Processing ELEC-E5550 Jan 30, 2019 Tiina Lindh-Knuutila D.Sc. (Tech) Lingsoft Language Services & Aalto University tiina.lindh-knuutila -at- aalto dot fi Material also by Mari-Sanna Paukkeri (mari-sanna.paukkeri -at- utopiaanalytics dot com) Today’s Agenda • Vector space models • word-document matrices • word vectors • stemming, weighting, dimensionality reduction • similarity measures • Count models vs. predictive models • Word2vec • Information retrieval (Briefly) • Course project details You shall know the word by the company it keeps • Language is symbolic in nature • Surface form is in an arbitrary relation with the meaning of the word • Hat vs. cat • One substitution: Levenshtein distance of 1 • Does not measure the semantic similarity of the words • Distributional semantics • Linguistic items which appear in similar contexts in large samples of language data tend to have similar meanings Firth, John R. 1957. A synopsis of linguistic theory 1930–1955. In Studies in linguistic analysis, 1–32. Oxford: Blackwell. George Miller and Walter Charles. 1991. Contextual correlates of semantic similarity. Language and Cognitive Processes, 6(1):1–28. Vector space models (VSM) • The use of a high-dimensional space of documents (or words) • Closeness in the vector space resembles closeness in the semantics or structure of the documents (depending on the features extracted). • Makes the use of data mining possible • Applications: – Document clustering/classification/… • Finding similar documents • Finding similar words – Word disambiguation – Information retrieval • Term discrimination: ranking keywords in the order of usefulness Vector space models (VSM) • Steps to build a vector space model 1. Preprocessing 2. -
Word Embeddings: History & Behind the Scene
Word Embeddings: History & Behind the Scene Alfan F. Wicaksono Information Retrieval Lab. Faculty of Computer Science Universitas Indonesia References • Bengio, Y., Ducharme, R., Vincent, P., & Janvin, C. (2003). A Neural Probabilistic Language Model. The Journal of Machine Learning Research, 3, 1137–1155. • Mikolov, T., Corrado, G., Chen, K., & Dean, J. (2013). Efficient Estimation of Word Representations in Vector Space. Proceedings of the International Conference on Learning Representations (ICLR 2013), 1–12 • Mikolov, T., Chen, K., Corrado, G., & Dean, J. (2013). Distributed Representations of Words and Phrases and their Compositionality. NIPS, 1–9. • Morin, F., & Bengio, Y. (2005). Hierarchical Probabilistic Neural Network Language Model. Aistats, 5. References Good weblogs for high-level understanding: • http://sebastianruder.com/word-embeddings-1/ • Sebastian Ruder. On word embeddings - Part 2: Approximating the Softmax. http://sebastianruder.com/word- embeddings-softmax • https://www.tensorflow.org/tutorials/word2vec • https://www.gavagai.se/blog/2015/09/30/a-brief-history-of- word-embeddings/ Some slides were also borrowed from From Dan Jurafsky’s course slide: Word Meaning and Similarity. Stanford University. Terminology • The term “Word Embedding” came from deep learning community • For computational linguistic community, they prefer “Distributional Semantic Model” • Other terms: – Distributed Representation – Semantic Vector Space – Word Space https://www.gavagai.se/blog/2015/09/30/a-brief-history-of-word-embeddings/ Before we learn Word Embeddings... Semantic Similarity • Word Similarity – Near-synonyms – “boat” and “ship”, “car” and “bicycle” • Word Relatedness – Can be related any way – Similar: “boat” and “ship” – Topical Similarity: • “boat” and “water” • “car” and “gasoline” Why Word Similarity? • Document Classification • Document Clustering • Language Modeling • Information Retrieval • .. -
Text Similarity Estimation Based on Word Embeddings and Matrix Norms for Targeted Marketing
Text Similarity Estimation Based on Word Embeddings and Matrix Norms for Targeted Marketing Tim vor der Bruck¨ Marc Pouly School of Information Technology School of Information Technology Lucerne University of Lucerne University of Applied Sciences and Arts Applied Sciences and Arts Switzerland Switzerland [email protected] [email protected] Abstract determined beforehand on a very large cor- pus typically using either the skip gram or The prevalent way to estimate the similarity of two documents based on word embeddings the continuous bag of words variant of the is to apply the cosine similarity measure to Word2Vec model (Mikolov et al., 2013). the two centroids obtained from the embed- 2. The centroid over all word embeddings be- ding vectors associated with the words in each longing to the same document is calculated document. Motivated by an industrial appli- to obtain its vector representation. cation from the domain of youth marketing, If vector representations of the two documents to where this approach produced only mediocre compare were successfully established, a similar- results, we propose an alternative way of com- ity estimate can be obtained by applying the cosine bining the word vectors using matrix norms. The evaluation shows superior results for most measure to the two vectors. of the investigated matrix norms in compari- Let x1; : : : ; xm and y1; : : : ; yn be the word vec- son to both the classical cosine measure and tors of two documents. The cosine similarity value several other document similarity estimates. between the two document centroids C1 und C2 is given by: 1 Introduction Estimating semantic document similarity is of ut- m n 1 X 1 X most importance in a lot of different areas, like cos( ( x ; y )) \ m i n i plagiarism detection, information retrieval, or text i=1 i=1 (1) Pm Pn summarization. -
UNIVERSITY of CALIFORNIA, SAN DIEGO Bag-Of-Concepts As a Movie
UNIVERSITY OF CALIFORNIA, SAN DIEGO Bag-of-Concepts as a Movie Genome and Representation A Thesis submitted in partial satisfaction of the requirements for the degree Master of Science in Computer Science by Colin Zhou Committee in charge: Professor Shlomo Dubnov, Chair Professor Sanjoy Dasgupta Professor Lawrence Saul 2016 The Thesis of Colin Zhou is approved, and it is acceptable in quality and form for publication on microfilm and electronically: Chair University of California, San Diego 2016 iii DEDICATION For Mom and Dad iv TABLE OF CONTENTS Signature Page........................................................................................................... iii Dedication.................................................................................................................. iv Table of Contents....................................................................................................... v List of Figures and Tables.......................................................................................... vi Abstract of the Thesis................................................................................................ vii Introduction................................................................................................................ 1 Chapter 1 Background............................................................................................... 3 1.1 Vector Space Models............................................................................... 3 1.1.1. Bag-of-Words........................................................................ -
Analysis of a Vector Space Model, Latent Semantic Indexing and Formal Concept Analysis for Information Retrieval
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES • Volume 12, No 1 Sofia • 2012 Analysis of a Vector Space Model, Latent Semantic Indexing and Formal Concept Analysis for Information Retrieval Ch. Aswani Kumar1, M. Radvansky2, J. Annapurna3 1School of Information Technology and Engineering, VIT University, Vellore, India 2VSB Technical University of Ostrava, Ostrava, Czech Republic 3School of Computing Science and Engineering, VIT University, Vellore, India Email: [email protected] Abstract: Latent Semantic Indexing (LSI), a variant of classical Vector Space Model (VSM), is an Information Retrieval (IR) model that attempts to capture the latent semantic relationship between the data items. Mathematical lattices, under the framework of Formal Concept Analysis (FCA), represent conceptual hierarchies in data and retrieve the information. However, both LSI and FCA use the data represented in the form of matrices. The objective of this paper is to systematically analyze VSM, LSI and FCA for the task of IR using standard and real life datasets. Keywords: Formal concept analysis, Information Retrieval, latent semantic indexing, vector space model. 1. Introduction Information Retrieval (IR) deals with the representation, storage, organization and access to information items. IR is highly iterative in human interactive process aimed at retrieving the documents that are related to users’ information needs. The human interaction consists in submitting information needed as a query, analyzing the ranked and retrieved documents, modifying the query and submitting iteratively until the actual documents related to the need are found or the search process is terminated by the user. Several techniques including a simple keyword to advanced NLP are available for developing IR systems. -
Using Semantic Folding with Textrank for Automatic Summarization
DEGREE PROJECT IN COMPUTER SCIENCE AND ENGINEERING, SECOND CYCLE, 30 CREDITS STOCKHOLM, SWEDEN 2017 Using semantic folding with TextRank for automatic summarization SIMON KARLSSON KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF COMPUTER SCIENCE AND COMMUNICATION Using semantic folding with TextRank for automatic summarization SIMON KARLSSON [email protected] Master in Computer Science Date: June 27, 2017 Principal: Findwise AB Supervisor at Findwise: Henrik Laurentz Supervisor at KTH: Stefan Nilsson Examiner at KTH: Olov Engwall Swedish title: TextRank med semantisk vikning för automatisk sammanfattning School of Computer Science and Communication i Abstract This master thesis deals with automatic summarization of text and how semantic fold- ing can be used as a similarity measure between sentences in the TextRank algorithm. The method was implemented and compared with two common similarity measures. These two similarity measures were cosine similarity of tf-idf vectors and the number of overlapping terms in two sentences. The three methods were implemented and the linguistic features used in the construc- tion were stop words, part-of-speech filtering and stemming. Five different part-of- speech filters were used, with different mixtures of nouns, verbs, and adjectives. The three methods were evaluated by summarizing documents from the Document Understanding Conference and comparing them to gold-standard summarization cre- ated by human judges. Comparison between the system summaries and gold-standard summaries was made with the ROUGE-1 measure. The algorithm with semantic fold- ing performed worst of the three methods, but only 0.0096 worse in F-score than cosine similarity of tf-idf vectors that performed best. For semantic folding, the average preci- sion was 46.2% and recall 45.7% for the best-performing part-of-speech filter. -
Lecture Note
Algorithms for Data Science: Lecture on Finding Similar Items Barna Saha 1 Finding Similar Items Finding similar items is a fundamental data mining task. We may want to find whether two documents are similar to detect plagiarism, mirror websites, multiple versions of the same article etc. Finding similar items is useful for building recommender systems as well where we want to find users with similar buying patterns. In Netflix two movies can be deemed similar if they are rated highly by the same customers. While, there are many measures of similarity, in this lecture, we will concentrate one such popular measure known as Jaccard Similarity. Definition (Jaccard Similairty). Given two sets S1 and S2, Jaccard similarity of S1 and S2 is defined as |S1∩S2 S1∪S2 Example 1. Let S1 = {1, 2, 3, 4, 7} and S2 = {1, 4, 9, 7, 5} then |S1 ∪ S2| = 7 and |S1 ∩ S2| = 3. 3 Thus the Jaccard similarity of S1 and S2 is 7 . 1.1 Document Similarity To compare two documents to know how similar they are, here is a simple approach: • Compute k shingles for a suitable value of k k shingles are all substrings of length k that appear in the document. For example, if a document is abcdabd and k = 2, then the 2-shingles are ab, bc, cd, da, ab, bd. • Compare the set of k shingled based on Jaccard similarity. One will often map the set of shingled to a set of integers via hashing. What should be the right size of the hash table? Select a hash table size to avoid collision.