Mapping Keywords to Linked Data Resources for Automatic Query Expansion
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
A Sentiment-Based Chat Bot
A sentiment-based chat bot Automatic Twitter replies with Python ALEXANDER BLOM SOFIE THORSEN Bachelor’s essay at CSC Supervisor: Anna Hjalmarsson Examiner: Mårten Björkman E-mail: [email protected], [email protected] Abstract Natural language processing is a field in computer science which involves making computers derive meaning from human language and input as a way of interacting with the real world. Broadly speaking, sentiment analysis is the act of determining the attitude of an author or speaker, with respect to a certain topic or the overall context and is an application of the natural language processing field. This essay discusses the implementation of a Twitter chat bot that uses natural language processing and sentiment analysis to construct a be- lievable reply. This is done in the Python programming language, using a statistical method called Naive Bayes classifying supplied by the NLTK Python package. The essay concludes that applying natural language processing and sentiment analysis in this isolated fashion was simple, but achieving more complex tasks greatly increases the difficulty. Referat Natural language processing är ett fält inom datavetenskap som innefat- tar att få datorer att förstå mänskligt språk och indata för att på så sätt kunna interagera med den riktiga världen. Sentiment analysis är, generellt sagt, akten av att försöka bestämma känslan hos en författare eller talare med avseende på ett specifikt ämne eller sammanhang och är en applicering av fältet natural language processing. Den här rapporten diskuterar implementeringen av en Twitter-chatbot som använder just natural language processing och sentiment analysis för att kunna svara på tweets genom att använda känslan i tweetet. -
Applying Ensembling Methods to BERT to Boost Model Performance
Applying Ensembling Methods to BERT to Boost Model Performance Charlie Xu, Solomon Barth, Zoe Solis Department of Computer Science Stanford University cxu2, sbarth, zoesolis @stanford.edu Abstract Due to its wide range of applications and difficulty to solve, Machine Comprehen- sion has become a significant point of focus in AI research. Using the SQuAD 2.0 dataset as a metric, BERT models have been able to achieve superior performance to most other models. Ensembling BERT with other models has yielded even greater results, since these ensemble models are able to utilize the relative strengths and adjust for the relative weaknesses of each of the individual submodels. In this project, we present various ensemble models that leveraged BiDAF and multiple different BERT models to achieve superior performance. We also studied the effects how dataset manipulations and different ensembling methods can affect the performance of our models. Lastly, we attempted to develop the Synthetic Self- Training Question Answering model. In the end, our best performing multi-BERT ensemble obtained an EM score of 73.576 and an F1 score of 76.721. 1 Introduction Machine Comprehension (MC) is a complex task in NLP that aims to understand written language. Question Answering (QA) is one of the major tasks within MC, requiring a model to provide an answer, given a contextual text passage and question. It has a wide variety of applications, including search engines and voice assistants, making it a popular problem for NLP researchers. The Stanford Question Answering Dataset (SQuAD) is a very popular means for training, tuning, testing, and comparing question answering systems. -
Aconcorde: Towards a Proper Concordance for Arabic
aConCorde: towards a proper concordance for Arabic Andrew Roberts, Dr Latifa Al-Sulaiti and Eric Atwell School of Computing University of Leeds LS2 9JT United Kingdom {andyr,latifa,eric}@comp.leeds.ac.uk July 12, 2005 Abstract Arabic corpus linguistics is currently enjoying a surge in activity. As the growth in the number of available Arabic corpora continues, there is an increased need for robust tools that can process this data, whether it be for research or teaching. One such tool that is useful for both groups is the concordancer — a simple tool for displaying a specified target word in its context. However, obtaining one that can reliably cope with the Arabic lan- guage had proved extremely difficult. Therefore, aConCorde was created to provide such a tool to the community. 1 Introduction A concordancer is a simple tool for summarising the contents of corpora based on words of interest to the user. Otherwise, manually navigating through (of- ten very large) corpora would be a long and tedious task. Concordancers are therefore extremely useful as a time-saving tool, but also much more. By isolat- ing keywords in their contexts, linguists can use concordance output to under- stand the behaviour of interesting words. Lexicographers can use the evidence to find if a word has multiple senses, and also towards defining their meaning. There is also much research and discussion about how concordance tools can be beneficial for data-driven language learning (Johns, 1990). A number of studies have demonstrated that providing access to corpora and concordancers benefited students learning a second language. -
Introduction to Linked Data and Its Lifecycle on the Web
Introduction to Linked Data and its Lifecycle on the Web Sören Auer, Jens Lehmann, Axel-Cyrille Ngonga Ngomo, Amrapali Zaveri AKSW, Institut für Informatik, Universität Leipzig, Pf 100920, 04009 Leipzig {lastname}@informatik.uni-leipzig.de http://aksw.org Abstract. With Linked Data, a very pragmatic approach towards achieving the vision of the Semantic Web has gained some traction in the last years. The term Linked Data refers to a set of best practices for publishing and interlinking struc- tured data on the Web. While many standards, methods and technologies devel- oped within by the Semantic Web community are applicable for Linked Data, there are also a number of specific characteristics of Linked Data, which have to be considered. In this article we introduce the main concepts of Linked Data. We present an overview of the Linked Data lifecycle and discuss individual ap- proaches as well as the state-of-the-art with regard to extraction, authoring, link- ing, enrichment as well as quality of Linked Data. We conclude the chapter with a discussion of issues, limitations and further research and development challenges of Linked Data. This article is an updated version of a similar lecture given at Reasoning Web Summer School 2011. 1 Introduction One of the biggest challenges in the area of intelligent information management is the exploitation of the Web as a platform for data and information integration as well as for search and querying. Just as we publish unstructured textual information on the Web as HTML pages and search such information by using keyword-based search engines, we are already able to easily publish structured information, reliably interlink this informa- tion with other data published on the Web and search the resulting data space by using more expressive querying beyond simple keyword searches. -
Linked Data Services for Internet of Things
International Conference on Recent Advances in Computer Systems (RACS 2015) Linked Data Services for Internet of Things Jaroslav Pullmann, Dr. Yehya Mohamad User-Centered Ubiquitous Computing department Fraunhofer Institute for Applied Information Technology FIT Sankt Augustin, Germany {jaroslav.pullmann, yehya.mohamad}@fit.fraunhofer.de Abstract — In this paper we present the open source “LinkSmart Resource Framework” allowing developers to II. LINKSMART RESOURCE FRAMEWORK incorporate heterogeneous data sources and physical devices through a generic service facade independent of the data model, A. Rationale persistence technology or access protocol. We particularly consi- The Java Data Objects standard [2] specifies an interface to der the integration and maintenance of Linked Data, a widely persist Java objects in a technology agnostic way. The related accepted means for expressing highly-structured, machine reada- ble (meta)data graphs. Thanks to its uniform, technology- Java Persistence specification [3] concentrates on object-rela- agnostic view on data, the framework is expected to increase the tional mapping only. We consider both specifications techno- ease-of-use, maintainability and usability of software relying on logy-driven, overly detailed with regards to common data ma- it. The development of various technologies to access, interact nagement tasks, while missing some important high-level with and manage data has led to rise of parallel communities functionality. The rationale underlying the LinkSmart Resour- often unaware of alternatives beyond their technology bounda- ce Platform is to define a generic, uniform interface for ries. Systems for object-relational mapping, NoSQL document or management, retrieval and processing of data while graph-based Linked Data storage usually exhibit complex, maintaining a technology and implementation agnostic facade. -
Using Shape Expressions (Shex) to Share RDF Data Models and to Guide Curation with Rigorous Validation B Katherine Thornton1( ), Harold Solbrig2, Gregory S
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repositorio Institucional de la Universidad de Oviedo Using Shape Expressions (ShEx) to Share RDF Data Models and to Guide Curation with Rigorous Validation B Katherine Thornton1( ), Harold Solbrig2, Gregory S. Stupp3, Jose Emilio Labra Gayo4, Daniel Mietchen5, Eric Prud’hommeaux6, and Andra Waagmeester7 1 Yale University, New Haven, CT, USA [email protected] 2 Johns Hopkins University, Baltimore, MD, USA [email protected] 3 The Scripps Research Institute, San Diego, CA, USA [email protected] 4 University of Oviedo, Oviedo, Spain [email protected] 5 Data Science Institute, University of Virginia, Charlottesville, VA, USA [email protected] 6 World Wide Web Consortium (W3C), MIT, Cambridge, MA, USA [email protected] 7 Micelio, Antwerpen, Belgium [email protected] Abstract. We discuss Shape Expressions (ShEx), a concise, formal, modeling and validation language for RDF structures. For instance, a Shape Expression could prescribe that subjects in a given RDF graph that fall into the shape “Paper” are expected to have a section called “Abstract”, and any ShEx implementation can confirm whether that is indeed the case for all such subjects within a given graph or subgraph. There are currently five actively maintained ShEx implementations. We discuss how we use the JavaScript, Scala and Python implementa- tions in RDF data validation workflows in distinct, applied contexts. We present examples of how ShEx can be used to model and validate data from two different sources, the domain-specific Fast Healthcare Interop- erability Resources (FHIR) and the domain-generic Wikidata knowledge base, which is the linked database built and maintained by the Wikimedia Foundation as a sister project to Wikipedia. -
Hunspell – the Free Spelling Checker
Hunspell – The free spelling checker About Hunspell Hunspell is a spell checker and morphological analyzer library and program designed for languages with rich morphology and complex word compounding or character encoding. Hunspell interfaces: Ispell-like terminal interface using Curses library, Ispell pipe interface, OpenOffice.org UNO module. Main features of Hunspell spell checker and morphological analyzer: - Unicode support (affix rules work only with the first 65535 Unicode characters) - Morphological analysis (in custom item and arrangement style) and stemming - Max. 65535 affix classes and twofold affix stripping (for agglutinative languages, like Azeri, Basque, Estonian, Finnish, Hungarian, Turkish, etc.) - Support complex compoundings (for example, Hungarian and German) - Support language specific features (for example, special casing of Azeri and Turkish dotted i, or German sharp s) - Handle conditional affixes, circumfixes, fogemorphemes, forbidden words, pseudoroots and homonyms. - Free software (LGPL, GPL, MPL tri-license) Usage The src/tools dictionary contains ten executables after compiling (or some of them are in the src/win_api): affixcompress: dictionary generation from large (millions of words) vocabularies analyze: example of spell checking, stemming and morphological analysis chmorph: example of automatic morphological generation and conversion example: example of spell checking and suggestion hunspell: main program for spell checking and others (see manual) hunzip: decompressor of hzip format hzip: compressor of -
Package 'Ptstem'
Package ‘ptstem’ January 2, 2019 Type Package Title Stemming Algorithms for the Portuguese Language Version 0.0.4 Description Wraps a collection of stemming algorithms for the Portuguese Language. URL https://github.com/dfalbel/ptstem License MIT + file LICENSE LazyData TRUE RoxygenNote 5.0.1 Encoding UTF-8 Imports dplyr, hunspell, magrittr, rslp, SnowballC, stringr, tidyr, tokenizers Suggests testthat, covr, plyr, knitr, rmarkdown VignetteBuilder knitr NeedsCompilation no Author Daniel Falbel [aut, cre] Maintainer Daniel Falbel <[email protected]> Repository CRAN Date/Publication 2019-01-02 14:40:02 UTC R topics documented: complete_stems . .2 extract_words . .2 overstemming_index . .3 performance . .3 ptstem_words . .4 stem_hunspell . .5 stem_modified_hunspell . .5 stem_porter . .6 1 2 extract_words stem_rslp . .7 understemming_index . .7 unify_stems . .8 Index 9 complete_stems Complete Stems Description Complete Stems Usage complete_stems(words, stems) Arguments words character vector of words stems character vector of stems extract_words Extract words Extracts all words from a character string of texts. Description Extract words Extracts all words from a character string of texts. Usage extract_words(texts) Arguments texts character vector of texts Note it uses the regex \b[:alpha:]+\b to extract words. overstemming_index 3 overstemming_index Overstemming Index (OI) Description It calculates the proportion of unrelated words that were combined. Usage overstemming_index(words, stems) Arguments words is a data.frame containing a column word a a column group so the function can identify groups of words. stems is a character vector with the stemming result for each word performance Performance Description Performance Usage performance(stemmers = c("rslp", "hunspell", "porter", "modified-hunspell")) Arguments stemmers a character vector with names of stemming algorithms. -
Cross-Lingual and Supervised Learning Approach for Indonesian Word Sense Disambiguation Task
Cross-Lingual and Supervised Learning Approach for Indonesian Word Sense Disambiguation Task Rahmad Mahendra, Heninggar Septiantri, Haryo Akbarianto Wibowo, Ruli Manurung, and Mirna Adriani Faculty of Computer Science, Universitas Indonesia Depok 16424, West Java, Indonesia [email protected], fheninggar, [email protected] Abstract after the target word), the nearest verb from the target word, the transitive verb around the target Ambiguity is a problem we frequently face word, and the document context. Unfortunately, in Natural Language Processing. Word neither the model nor the corpus from this research Sense Disambiguation (WSD) is a task to is made publicly available. determine the correct sense of an ambigu- This paper reports our study on WSD task for ous word. However, research in WSD for Indonesian using the combination of the cross- Indonesian is still rare to find. The avail- lingual and supervised learning approach. Train- ability of English-Indonesian parallel cor- ing data is automatically acquired using Cross- pora and WordNet for both languages can Lingual WSD (CLWSD) approach by utilizing be used as training data for WSD by ap- WordNet and parallel corpus. Then, the monolin- plying Cross-Lingual WSD method. This gual WSD model is built from training data and training data is used as an input to build it is used to assign the correct sense to any previ- a model using supervised machine learn- ously unseen word in a new context. ing algorithms. Our research also exam- ines the use of Word Embedding features 2 Related Work to build the WSD model. WSD task is undertaken in two main steps, namely 1 Introduction listing all possible senses for a word and deter- mining the right sense given its context (Ide and One of the biggest challenges in Natural Language Veronis,´ 1998). -
A Comparison of Word Embeddings and N-Gram Models for Dbpedia Type and Invalid Entity Detection †
information Article A Comparison of Word Embeddings and N-gram Models for DBpedia Type and Invalid Entity Detection † Hanqing Zhou *, Amal Zouaq and Diana Inkpen School of Electrical Engineering and Computer Science, University of Ottawa, Ottawa ON K1N 6N5, Canada; [email protected] (A.Z.); [email protected] (D.I.) * Correspondence: [email protected]; Tel.: +1-613-562-5800 † This paper is an extended version of our conference paper: Hanqing Zhou, Amal Zouaq, and Diana Inkpen. DBpedia Entity Type Detection using Entity Embeddings and N-Gram Models. In Proceedings of the International Conference on Knowledge Engineering and Semantic Web (KESW 2017), Szczecin, Poland, 8–10 November 2017, pp. 309–322. Received: 6 November 2018; Accepted: 20 December 2018; Published: 25 December 2018 Abstract: This article presents and evaluates a method for the detection of DBpedia types and entities that can be used for knowledge base completion and maintenance. This method compares entity embeddings with traditional N-gram models coupled with clustering and classification. We tackle two challenges: (a) the detection of entity types, which can be used to detect invalid DBpedia types and assign DBpedia types for type-less entities; and (b) the detection of invalid entities in the resource description of a DBpedia entity. Our results show that entity embeddings outperform n-gram models for type and entity detection and can contribute to the improvement of DBpedia’s quality, maintenance, and evolution. Keywords: semantic web; DBpedia; entity embedding; n-grams; type identification; entity identification; data mining; machine learning 1. Introduction The Semantic Web is defined by Berners-Lee et al. -
Linked Data Vs Schema.Org: a Town Hall Debate About the Future of Information
Linked Data vs Schema.org: A Town Hall Debate about the Future of Information Schema.org Resources Schema.org Microdata Creation and Analysis Tools Schema.org Google Structured Data Testing Tool http://schema.org/docs/full.html http://www.google.com/webmasters/tools/richsnippets The entire hierarchy in one file. Microdata Parser Schema.org FAQ http://tools.seomoves.org/microdata http://schema.org/docs/faq.html This tool parses HTML5 microdata on a web page and displays the results in a graphical view. You can see the full details of each microdata type by clicking on it. Getting Started with Schema.org http://schema.org/docs/gs.html Drupal module Includes information about how to mark up your content http://drupal.org/project/schemaorg using microdata, using the schema.org vocabulary, and advanced topics. A drop-in solution to enable the collections of schemas available at schema.org on your Drupal 7 site. Micro Data & Schema.org: Guide To Generating Rich Snippets Wordpress plugins http://seogadget.com/micro-data-schema-org-guide-to- http://wordpress.org/extend/plugins/tags/schemaorg generating-rich-snippets A list of Wordpress plugins that allow you to easily insert A very comprehensive guide that includes an schema.org microdata on your site. introduction, a section on integrating microdata (use cases) schema.org, and a section on tools and useful Microdata generator resources. http://www.microdatagenerator.com A simple generator that allows you to input basic HTML5 Microdata and Schema.org information and have that info converted into the http://journal.code4lib.org/articles/6400 standard schema.org markup structure. -
FAIR Linked Data
FAIR Linked Data - Towards a Linked Data backbone for users and machines Johannes Frey Sebastian Hellmann [email protected] Knowledge Integration and Linked Data Technologies (KILT/AKSW) & DBpedia Association InfAI, Leipzig University Leipzig, Germany ABSTRACT scientific and open data. While the principles became widely ac- Although many FAIR principles could be fulfilled by 5-star Linked cepted and popular, they lack technical details and clear, measurable Open Data, the successful realization of FAIR poses a multitude criteria. of challenges. FAIR publishing and retrieval of Linked Data is still Three years ago a dedicated European Commission Expert Group rather a FAIRytale than reality, for users and machines. In this released an action plan for "turning FAIR into reality" [1] empha- paper, we give an overview on four major approaches that tackle sizing the need for defining "FAIR for implementation". individual challenges of FAIR data and present our vision of a FAIR In this paper, we argue that Linked Data as a technology in Linked Data backbone. We propose 1) DBpedia Databus - a flex- combination with the Linked Open Data movement and ontologies ible, heavily automatizable dataset management and publishing can be seen as one of the most promising approaches for managing platform based on DataID metadata; that is extended by 2) the scholarly metadata as well as representing and connecting research novel Databus Mods architecture which allows for flexible, uni- results (RDF knowledge graphs) across the web. Although many fied, community-specific metadata extensions and (search) overlay FAIR principles could be fulfilled by 5-star Linked Open Data [5], systems; 3) DBpedia Archivo an archiving solution for unified han- the successful realization of FAIR principles for the Web of Data in dling and improvement of FAIRness for ontologies on publisher and its current state is a FAIRytale.