Adaptive Semantic Annotation of Entity and Concept Mentionsin Text

Total Page:16

File Type:pdf, Size:1020Kb

Adaptive Semantic Annotation of Entity and Concept Mentionsin Text i ADAPTIVE SEMANTIC ANNOTATION OF ENTITY AND CONCEPT MENTIONS IN TEXT A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy By PABLO N. MENDES M.Sc., Federal University of Rio de Janeiro, 2005 B.Sc., Federal University of Juiz de Fora, 2003 2013 Wright State University Wright State University GRADUATE SCHOOL June 1, 2014 I HEREBY RECOMMEND THAT THE THESIS PREPARED UNDER MY SUPER- VISION BY Pablo N. Mendes ENTITLED Adaptive Semantic Annotation of Entity and Concept Mentions in Text BE ACCEPTED IN PARTIAL FULFILLMENT OF THE RE- QUIREMENTS FOR THE DEGREE OF Doctor of Philosophy. Amit P. Sheth, Ph.D. Thesis Director Arthur Goshtasby, Ph.D. Director, Computer Science & Engineering Graduate Program R. William Ayres, Ph.D. Interim Dean, Graduate School Committee on Final Examination Amit P. Sheth, Ph.D. Krishnaprasad Thirunarayan, Ph.D. Shajoun Wang, Ph.D. Soren¨ Auer, Ph.D. ABSTRACT Mendes, Pablo N. Ph.D., Department of Computer Science and Engineering, Wright State Univer- sity, 2013. Adaptive Semantic Annotation of Entity and Concept Mentions in Text. The recent years have seen an increase in interest for knowledge repositories that are useful across applications, in contrast to the creation of ad hoc or application-specific databases. These knowledge repositories figure as a central provider of unambiguous iden- tifiers and semantic relationships between entities. As such, these shared entity descriptions serve as a common vocabulary to exchange and organize information in different formats and for different purposes. Therefore, there has been remarkable interest in systems that are able to automatically tag textual documents with identifiers from shared knowledge repositories so that the content in those documents is described in a vocabulary that is unambiguously understood across applications. Tagging textual documents according to these knowledge bases is a challenging task. It involves recognizing the entities and concepts that have been mentioned in a particular passage and attempting to resolve eventual ambiguity of language in order to choose one of many possible meanings for a phrase. There has been substantial work on recognizing and disambiguating entities for specialized applications, or constrained to limited entity types and particular types of text. In the context of shared knowledge bases, since each appli- cation has potentially very different needs, systems must have unprecedented breadth and flexibility to ensure their usefulness across applications. Documents may exhibit different language and discourse characteristics, discuss very diverse topics, or require the focus on parts of the knowledge repository that are inherently harder to disambiguate. In practice, for developers looking for a system to support their use case, is often unclear if an exist- ing solution is applicable, leading those developers to trial-and-error and ad hoc usage of multiple systems in an attempt to achieve their objective. In this dissertation, I propose a conceptual model that unifies related techniques in iii this space under a common multi-dimensional framework that enables the elucidation of strengths and limitations of each technique, supporting developers in their search for a suitable tool for their needs. Moreover, the model serves as the basis for the development of flexible systems that have the ability of supporting document tagging for different use cases. I describe such an implementation, DBpedia Spotlight, along with extensions that we performed to the knowledge base DBpedia to support this implementation. I report evaluations of this tool on several well known data sets, and demonstrate applications to diverse use cases for further validation. iv Contents 1 Introduction1 1.1 Historical Context...............................1 1.2 Use Cases for Semantic Annotation.....................3 1.3 Thesis statement................................5 1.4 Organization..................................6 1.5 Notation....................................6 2 Background and Related Work7 2.1 Methodology for the Literature Review...................7 2.2 A Review of Information Extraction Tasks..................8 2.3 State of the art................................. 17 2.4 Cross-task Benchmarks............................ 18 2.5 Conclusion.................................. 20 3 A Conceptual Framework for Semantic Annotation 21 3.1 Definitions................................... 21 3.1.1 Actors................................. 22 3.1.2 Objective............................... 25 3.1.3 Textual Content............................ 26 3.1.4 Knowledge Base........................... 27 3.1.5 System................................ 29 3.1.6 Annotations.............................. 34 3.2 Conclusion.................................. 35 4 The DBpedia Knowledge Base 36 4.1 Extracting a Knowledge Graph from Wikipedia............... 36 4.1.1 Extracting Triples........................... 37 4.1.2 The DBpedia Ontology........................ 38 4.1.3 Cross-Language Data Fusion..................... 39 4.1.4 RDF Links to other Data Sets.................... 42 4.2 Supporting Natural Language Processing.................. 43 4.2.1 The Lexicalization Data Set..................... 43 v 4.2.2 The Thematic Concepts Data Set................... 45 4.2.3 The Grammatical Gender Data Set.................. 46 4.2.4 Occurrence Statistics Data Set.................... 47 4.2.5 The Topic Signatures Data Set.................... 47 4.3 Conclusion.................................. 48 5 DBpedia Spotlight: A System for Adaptive Knowledge Base Tagging of DBpe- dia Entities and Concepts 49 5.1 Implementation................................ 50 5.1.1 Phrase Recognition.......................... 52 5.1.2 Candidate Selection......................... 56 5.1.3 Disambiguation............................ 56 5.1.4 Relatedness.............................. 59 5.1.5 Tagging................................ 59 5.2 Using DBpedia Spotlight........................... 61 5.2.1 Web Application........................... 62 5.2.2 Web Service............................. 62 5.2.3 Installation.............................. 64 5.3 Continuous Evolution............................. 65 5.3.1 Live updates............................. 65 5.3.2 Feedback incorporation........................ 66 5.3.3 Round-trip semantics......................... 66 5.4 Conclusion.................................. 68 6 Core Evaluations 69 6.1 Evaluation Corpora.............................. 69 6.1.1 Wikipedia............................... 69 6.1.2 CSAW................................ 69 6.2 Spotting Evaluation Results.......................... 70 6.2.1 Error Analysis............................ 72 6.3 Disambiguation Evaluation Results...................... 72 6.3.1 TAC-KBP............................... 74 6.4 Annotation Evaluation Results........................ 76 6.4.1 News Articles............................. 76 6.5 A Framework for Evaluating Difficulty to Disambiguate.......... 79 6.5.1 Comparison with annotation-as-a-service systems.......... 81 6.6 Conclusions.................................. 84 7 Case Studies 85 7.1 Named Entity Recognition in Tweets..................... 85 7.2 Managing Information Overload....................... 88 7.3 Understanding Website Similarity...................... 92 7.3.1 Entity-aware Click Graph...................... 93 7.3.2 Website Similarity Graph....................... 95 7.3.3 Results................................ 95 vi 7.4 Automated Audio Tagging.......................... 96 7.5 Educational Material - Emergency Management............... 97 8 Conclusion 100 8.1 Limitations.................................. 101 Bibliography 102 vii List of Figures 1.1 Example SPO triples describing the Brazilian city of Rio de Janeiro (iden- tified by the URI Rio_de_Janeiro). It describes Rio’s total area in square miles through a property (areaTotalSqMi) and relates it to an- other resource (Brazil) through the property country. Namespaces have been omitted for readability........................2 2.1 Key terms used in the literature review.....................8 3.1 Workflow of a Knowledge Base Tagging (KBT) task............. 22 3.2 Interactions between actors and a KBT system................ 23 3.3 Typical internal workflow of a KBT system.................. 30 4.1 A snippet of the triples describing dbpedia:Rio_de_Janeiro..... 37 4.2 Linking Open Data Cloud diagram [Jentzsch et al., 2011]........... 42 4.3 SPARQL query demonstrating how to retrieve entities and concepts under a certain category................................ 45 4.4 SPARQL query demonstrating how to retrieve pages linking to topical con- cepts...................................... 46 4.5 A snippet of the Grammatical Gender Data Set................ 46 4.6 A snippet of the Topic Signatures Data Set.................. 48 5.1 Model objects in DBpedia Spotlight’s core model............... 51 5.2 Example phrases recognized externally and encoded as SpotXml....... 55 5.3 Results for a request to /related for the top-ranked URIs by related- ness to dbpedia:Adidas. Each key-value pair represents a URI and the relatedness score (TF*IDF) between that URI and Adidas.......... 59 5.4 DBpedia Spotlight Web Application...................... 62 5.5 Example call to the Web Service using cURL................. 63 5.6 Example XML fragment resulting from the annotation service........ 64 5.7 Round-trip Semantics............................. 67 3 3+ 6.1 Comparison of B and B F1 scores for NIL and Non-NIL
Recommended publications
  • Metadata for Semantic and Social Applications
    etadata is a key aspect of our evolving infrastructure for information management, social computing, and scientific collaboration. DC-2008M will focus on metadata challenges, solutions, and innovation in initiatives and activities underlying semantic and social applications. Metadata is part of the fabric of social computing, which includes the use of wikis, blogs, and tagging for collaboration and participation. Metadata also underlies the development of semantic applications, and the Semantic Web — the representation and integration of multimedia knowledge structures on the basis of semantic models. These two trends flow together in applications such as Wikipedia, where authors collectively create structured information that can be extracted and used to enhance access to and use of information sources. Recent discussion has focused on how existing bibliographic standards can be expressed as Semantic Metadata for Web vocabularies to facilitate the ingration of library and cultural heritage data with other types of data. Harnessing the efforts of content providers and end-users to link, tag, edit, and describe their Semantic and information in interoperable ways (”participatory metadata”) is a key step towards providing knowledge environments that are scalable, self-correcting, and evolvable. Social Applications DC-2008 will explore conceptual and practical issues in the development and deployment of semantic and social applications to meet the needs of specific communities of practice. Edited by Jane Greenberg and Wolfgang Klas DC-2008
    [Show full text]
  • A Sentiment-Based Chat Bot
    A sentiment-based chat bot Automatic Twitter replies with Python ALEXANDER BLOM SOFIE THORSEN Bachelor’s essay at CSC Supervisor: Anna Hjalmarsson Examiner: Mårten Björkman E-mail: [email protected], [email protected] Abstract Natural language processing is a field in computer science which involves making computers derive meaning from human language and input as a way of interacting with the real world. Broadly speaking, sentiment analysis is the act of determining the attitude of an author or speaker, with respect to a certain topic or the overall context and is an application of the natural language processing field. This essay discusses the implementation of a Twitter chat bot that uses natural language processing and sentiment analysis to construct a be- lievable reply. This is done in the Python programming language, using a statistical method called Naive Bayes classifying supplied by the NLTK Python package. The essay concludes that applying natural language processing and sentiment analysis in this isolated fashion was simple, but achieving more complex tasks greatly increases the difficulty. Referat Natural language processing är ett fält inom datavetenskap som innefat- tar att få datorer att förstå mänskligt språk och indata för att på så sätt kunna interagera med den riktiga världen. Sentiment analysis är, generellt sagt, akten av att försöka bestämma känslan hos en författare eller talare med avseende på ett specifikt ämne eller sammanhang och är en applicering av fältet natural language processing. Den här rapporten diskuterar implementeringen av en Twitter-chatbot som använder just natural language processing och sentiment analysis för att kunna svara på tweets genom att använda känslan i tweetet.
    [Show full text]
  • Applying Ensembling Methods to BERT to Boost Model Performance
    Applying Ensembling Methods to BERT to Boost Model Performance Charlie Xu, Solomon Barth, Zoe Solis Department of Computer Science Stanford University cxu2, sbarth, zoesolis @stanford.edu Abstract Due to its wide range of applications and difficulty to solve, Machine Comprehen- sion has become a significant point of focus in AI research. Using the SQuAD 2.0 dataset as a metric, BERT models have been able to achieve superior performance to most other models. Ensembling BERT with other models has yielded even greater results, since these ensemble models are able to utilize the relative strengths and adjust for the relative weaknesses of each of the individual submodels. In this project, we present various ensemble models that leveraged BiDAF and multiple different BERT models to achieve superior performance. We also studied the effects how dataset manipulations and different ensembling methods can affect the performance of our models. Lastly, we attempted to develop the Synthetic Self- Training Question Answering model. In the end, our best performing multi-BERT ensemble obtained an EM score of 73.576 and an F1 score of 76.721. 1 Introduction Machine Comprehension (MC) is a complex task in NLP that aims to understand written language. Question Answering (QA) is one of the major tasks within MC, requiring a model to provide an answer, given a contextual text passage and question. It has a wide variety of applications, including search engines and voice assistants, making it a popular problem for NLP researchers. The Stanford Question Answering Dataset (SQuAD) is a very popular means for training, tuning, testing, and comparing question answering systems.
    [Show full text]
  • From the Semantic Web to Social Machines
    ARTICLE IN PRESS ARTINT:2455 JID:ARTINT AID:2455 /REV [m3G; v 1.23; Prn:25/11/2009; 12:36] P.1 (1-6) Artificial Intelligence ••• (••••) •••–••• Contents lists available at ScienceDirect Artificial Intelligence www.elsevier.com/locate/artint From the Semantic Web to social machines: A research challenge for AI on the World Wide Web ∗ Jim Hendler a, , Tim Berners-Lee b a Tetherless World Constellation, RPI, United States b Computer Science and AI Laboratory, MIT, United States article info abstract Article history: The advent of social computing on the Web has led to a new generation of Web Received 24 September 2009 applications that are powerful and world-changing. However, we argue that we are just Received in revised form 1 October 2009 at the beginning of this age of “social machines” and that their continued evolution and Accepted 1 October 2009 growth requires the cooperation of Web and AI researchers. In this paper, we show how Available online xxxx the growing Semantic Web provides necessary support for these technologies, outline the challenges we see in bringing the technology to the next level, and propose some starting places for the research. © 2009 Elsevier B.V. All rights reserved. Much has been written about the profound impact that the World Wide Web has had on society. Yet it is primarily in the past few years, as more interactive “read/write” technologies (e.g. Wikis, blogs and photo/video sharing) and social network- ing sites have proliferated, that the truly profound nature of this change is being felt. From the very beginning, however, the Web was designed to create a network of humans changing society empowered using this shared infrastructure.
    [Show full text]
  • Wikimedia Conferentie Nederland 2012 Conferentieboek
    http://www.wikimediaconferentie.nl WCN 2012 Conferentieboek CC-BY-SA 9:30–9:40 Opening 9:45–10:30 Lydia Pintscher: Introduction to Wikidata — the next big thing for Wikipedia and the world Wikipedia in het onderwijs Technische kant van wiki’s Wiki-gemeenschappen 10:45–11:30 Jos Punie Ralf Lämmel Sarah Morassi & Ziko van Dijk Het gebruik van Wikimedia Commons en Wikibooks in Community and ontology support for the Wikimedia Nederland, wat is dat precies? interactieve onderwijsvormen voor het secundaire onderwijs 101wiki 11:45–12:30 Tim Ruijters Sandra Fauconnier Een passie voor leren, de Nederlandse wikiversiteit Projecten van Wikimedia Nederland in 2012 en verder Bliksemsessie …+discussie …+vragensessie Lunch 13:15–14:15 Jimmy Wales 14:30–14:50 Wim Muskee Amit Bronner Lotte Belice Baltussen Wikipedia in Edurep Bridging the Gap of Multilingual Diversity Open Cultuur Data: Een bottom-up initiatief vanuit de erfgoedsector 14:55–15:15 Teun Lucassen Gerard Kuys Finne Boonen Scholieren op Wikipedia Onderwerpen vinden met DBpedia Blijf je of ga je weg? 15:30–15:50 Laura van Broekhoven & Jan Auke Brink Jeroen De Dauw Jan-Bart de Vreede 15:55–16:15 Wetenschappelijke stagiairs vertalen onderzoek naar Structured Data in MediaWiki Wikiwijs in vergelijking tot Wikiversity en Wikibooks Wikipedia–lemma 16:20–17:15 Prijsuitreiking van Wiki Loves Monuments Nederland 17:20–17:30 Afsluiting 17:30–18:30 Borrel Inhoudsopgave Organisatie 2 Voorwoord 3 09:45{10:30: Lydia Pintscher 4 13:15{14:15: Jimmy Wales 4 Wikipedia in het onderwijs 5 11:00{11:45: Jos Punie
    [Show full text]
  • EVERYTHING YOU NEED to KNOW ABOUT SEMANTIC SEO by GENNARO CUOFANO, ANDREA VOLPINI, and MARIA SILVIA SANNA the Semantic Web Is Here
    WHITE PAPER EVERYTHING YOU NEED TO KNOW ABOUT SEMANTIC SEO BY GENNARO CUOFANO, ANDREA VOLPINI, AND MARIA SILVIA SANNA The Semantic Web is here. Those that are taking advantage of Semantic Technologies to build a Semantic SEO strategy are benefiting from staggering results. From a research paper put together with the team atWordLift , presented at SEMANTiCS 2017, we documented that structured data is compelling from the digital marketing standpoint. For instance, on the analysis of the design-focused website freeyork.org, after three months of using structured data in their WordPress website we saw the following improvements: • +12.13% new users • +18.47% increase in organic traffic • +2.4 times increase in page views • +13.75% of sessions duration In other words, many still think of Semantic Technologies belonging to the future, when in reality quite a few players in the digital marketing space are taking advantage of them already. Semantic SEO is a new and powerful way to make your content strategy more effective. In this article/guide I will explain from scratch what Semantic SEO is and why it’s important. Why Semantic SEO? In a nutshell, search engines need context to understand a query properly and to fetch relevant results for it. Contexts are built using words, expressions, and other combinations of words and links as they appear in bodies of knowledge such as encyclopedias and large corpora of text. Semantic SEO is a marketing technique that improves the traffic of a website byproviding meaningful data that can unambiguously answer a specific search intent. It is also a way to create clusters of content that are semantically grouped into topics rather than keywords.
    [Show full text]
  • Towards a Korean Dbpedia and an Approach for Complementing the Korean Wikipedia Based on Dbpedia
    Towards a Korean DBpedia and an Approach for Complementing the Korean Wikipedia based on DBpedia Eun-kyung Kim1, Matthias Weidl2, Key-Sun Choi1, S¨orenAuer2 1 Semantic Web Research Center, CS Department, KAIST, Korea, 305-701 2 Universit¨at Leipzig, Department of Computer Science, Johannisgasse 26, D-04103 Leipzig, Germany [email protected], [email protected] [email protected], [email protected] Abstract. In the first part of this paper we report about experiences when applying the DBpedia extraction framework to the Korean Wikipedia. We improved the extraction of non-Latin characters and extended the framework with pluggable internationalization components in order to fa- cilitate the extraction of localized information. With these improvements we almost doubled the amount of extracted triples. We also will present the results of the extraction for Korean. In the second part, we present a conceptual study aimed at understanding the impact of international resource synchronization in DBpedia. In the absence of any informa- tion synchronization, each country would construct its own datasets and manage it from its users. Moreover the cooperation across the various countries is adversely affected. Keywords: Synchronization, Wikipedia, DBpedia, Multi-lingual 1 Introduction Wikipedia is the largest encyclopedia of mankind and is written collaboratively by people all around the world. Everybody can access this knowledge as well as add and edit articles. Right now Wikipedia is available in 260 languages and the quality of the articles reached a high level [1]. However, Wikipedia only offers full-text search for this textual information. For that reason, different projects have been started to convert this information into structured knowledge, which can be used by Semantic Web technologies to ask sophisticated queries against Wikipedia.
    [Show full text]
  • A New Generation Digital Content Service for Cultural Heritage Institutions
    A New Generation Digital Content Service for Cultural Heritage Institutions Pierfrancesco Bellini, Ivan Bruno, Daniele Cenni, Paolo Nesi, Michela Paolucci, and Marco Serena Distributed Systems and Internet Technology Lab, DISIT, Dipartimento di Ingegneria dell’Informazione, Università degli Studi di Firenze, Italy [email protected] http://www.disit.dsi.unifi.it Abstract. The evolution of semantic technology and related impact on internet services and solutions, such as social media, mobile technologies, etc., have de- termined a strong evolution in digital content services. Traditional content based online services are leaving the space to a new generation of solutions. In this paper, the experience of one of those new generation digital content service is presented, namely ECLAP (European Collected Library of Artistic Perform- ance, http://www.eclap.eu). It has been partially founded by the European Commission and includes/aggregates more than 35 international institutions. ECLAP provides services and tools for content management and user network- ing. They are based on a set of newly researched technologies and features in the area of semantic computing technologies capable of mining and establishing relationships among content elements, concepts and users. On this regard, ECLAP is a place in which these new solutions are made available for inter- ested institutions. Keywords: best practice network, semantic computing, recommendations, automated content management, content aggregation, social media. 1 Introduction Traditional library services in which the users can access to content by searching and browsing on-line catalogues obtaining lists of references and sporadically digital items (documents, images, etc.) are part of our history. With the introduction of web2.0/3.0, and thus of data mining and semantic computing, including social media and mobile technologies most of the digital libraries and museum services became rapidly obsolete and were constrained to rapidly change.
    [Show full text]
  • Aconcorde: Towards a Proper Concordance for Arabic
    aConCorde: towards a proper concordance for Arabic Andrew Roberts, Dr Latifa Al-Sulaiti and Eric Atwell School of Computing University of Leeds LS2 9JT United Kingdom {andyr,latifa,eric}@comp.leeds.ac.uk July 12, 2005 Abstract Arabic corpus linguistics is currently enjoying a surge in activity. As the growth in the number of available Arabic corpora continues, there is an increased need for robust tools that can process this data, whether it be for research or teaching. One such tool that is useful for both groups is the concordancer — a simple tool for displaying a specified target word in its context. However, obtaining one that can reliably cope with the Arabic lan- guage had proved extremely difficult. Therefore, aConCorde was created to provide such a tool to the community. 1 Introduction A concordancer is a simple tool for summarising the contents of corpora based on words of interest to the user. Otherwise, manually navigating through (of- ten very large) corpora would be a long and tedious task. Concordancers are therefore extremely useful as a time-saving tool, but also much more. By isolat- ing keywords in their contexts, linguists can use concordance output to under- stand the behaviour of interesting words. Lexicographers can use the evidence to find if a word has multiple senses, and also towards defining their meaning. There is also much research and discussion about how concordance tools can be beneficial for data-driven language learning (Johns, 1990). A number of studies have demonstrated that providing access to corpora and concordancers benefited students learning a second language.
    [Show full text]
  • On Using Product-Specific Schema.Org from Web Data Commons: an Empirical Set of Best Practices
    On using Product-Specific Schema.org from Web Data Commons: An Empirical Set of Best Practices Ravi Kiran Selvam Mayank Kejriwal [email protected] [email protected] USC Information Sciences Institute USC Information Sciences Institute Marina del Rey, California Marina del Rey, California ABSTRACT both to advertise and find products on the Web, in no small part Schema.org has experienced high growth in recent years. Struc- due to the ability of search engines to make good use of this data. tured descriptions of products embedded in HTML pages are now There is also limited evidence that structured data plays a key role not uncommon, especially on e-commerce websites. The Web Data in populating Web-scale knowledge graphs such as the Google Commons (WDC) project has extracted schema.org data at scale Knowledge Graph that are essential to modern semantic search from webpages in the Common Crawl and made it available as an [17]. RDF ‘knowledge graph’ at scale. The portion of this data that specif- However, even though most e-commerce platforms have their ically describes products offers a golden opportunity for researchers own proprietary datasets, some resources do exist for smaller com- and small companies to leverage it for analytics and downstream panies, researchers and organizations. A particularly important applications. Yet, because of the broad and expansive scope of this source of data is schema.org markup in webpages. Launched in data, it is not evident whether the data is usable in its raw form. In the early 2010s by major search engines such as Google and Bing, this paper, we do a detailed empirical study on the product-specific schema.org was designed to facilitate structured (and even knowl- schema.org data made available by WDC.
    [Show full text]
  • Implementation of Intelligent Semantic Web Search Engine
    International Journal of Engineering Research & Technology (IJERT) ISSN: 2278-0181 Vol. 4 Issue 04, April-2015 Implementation of Intelligent Semantic Web Search Engine Dinesh Jagtap*1, Nilesh Argade1, Shivaji Date1, Sainath Hole1, Mahendra Salunke2 1Department of Computer Engineering, 2Associate Professor, Department of Computer Engineering Sinhgad Institute of Technology and Science, Pune, Maharashtra, India 411041. II. BACKGROUND Abstract—Search engines are design for to search particular The idea of search engine and info retrieval from search information for a large database that is from World Wide engine is not a new concept. The interesting thing about Web. There are lots of search engines available. Google, traditional search engine is that different search engine will yahoo, Bing are the search engines which are most widely provide different result for the same query. While used search engines in today. The main objective of any search engines is to provide particular or required information was available in web, we have some fields of information with minimum time. The semantics web search problem in search engines. Information retrieval by engines are the next version of traditional search engines. The searching information on the web is not a fresh idea but has main problem of traditional search engines is that different challenges when it is compared to general information retrieval from the database is difficult or takes information retrieval. Different search engines return long time. Hence efficiency of search engines is reduced. To different search results due to the variation in indexing and overcome this intelligent semantic search engines are search process. Google, Yahoo, and Bing have been out introduced.
    [Show full text]
  • Chaudron: Extending Dbpedia with Measurement Julien Subercaze
    Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue.
    [Show full text]