Lookup, Explore, Discover: How Dbpedia Can Improve Your Web Search

Total Page:16

File Type:pdf, Size:1020Kb

Lookup, Explore, Discover: How Dbpedia Can Improve Your Web Search Undefined 0 (2010) 1{0 1 IOS Press Lookup, Explore, Discover: how DBpedia can improve your Web search Roberto Mirizzi a, Azzurra Ragone a Tommaso Di Noia a and Eugenio Di Sciascio a a Politecnico di Bari, via Re David 200, 7015, Bari { Italy E-mail: [email protected], [email protected], [email protected], [email protected] Abstract. The power of search is with no doubt one of the main aspects for the success of the Web. Differently from traditional searches in a library, search engines for digital libraries (as the Web is) allow to return results that are considerably effective. Nevertheless, current search engines are no more able to really help the user in all those tasks that go under the umbrella of exploratory search, where the user needs not only to perform a lookup operation but also to discover, understand and learn novel contents on complex topics while searching. Similar problems arise also with search exploiting classical keyword-based tagging techniques. In the paper we present LED, a web based system that aims to improve Web search by enabling users to properly explore knowledge associated to a query/keyword. We rely on DBpedia to explore the semantics of keywords and to guide the user during the exploration of search results. 1. Introduction happens when the user has not a clear idea of what she is looking for and (possibly) she needs some Thanks to the social paradigm of the Web 2.0 suggestions in the exploration process. The above new tools and technologies emerged to allow the mentioned tools enrich the page of search results user to interact with on-line resources. By now, we from classical search engines, e.g., Google, with a are very familiar with the notion of annotating a tag cloud suggesting relevant keywords, so that the web page as well as of tagging system, tag cloud, user may expand her query by adding new terms tag browsing just to cite a few. Surfing the wave of just clicking on tags in the cloud. the social Web, several web search tools have been From the users perspective, a very interesting developed to help users to navigate through search aspect related to the generation of the tag cloud results in a fast, simpler and efficient way combin- in the page of search results is the combination ing traditional search facilities with the tag cloud of lookup search with exploratory browsing. As exploration. Such tools are usually available as pointed out by Marchionini in [8] there are mainly plug-ins (add-ons) for the most common browsers two different ways of search strategies: lookup and and the most relevant are currently: DeeperWeb1, exploratory search. The presentation of contex- Search Cloudlet 2, SenseBot - Search Results Sum- tualized tag clouds combined with search results marizer 3. The main aim of these plug-ins is to makes possible to combine the two search strate- guide users in refining queries with terms which gies. As a matter of fact, after a lookup in a search are somehow related with the ones they are look- engine the user may explore the results space. ing thus exploring the results space. This usually Basically, all the above cited tools exploit text retrieval techniques to suggest new tags looking for relevant keywords in the snippets of search en- 1http://www.deeperweb.com 2http://www.getcloudlet.com gines results. Terms in the tag cloud are the most 3http://www.sensebot.net frequent ones in the result snippets. None of these 0000-0000/10/$00.00 c 2010 { IOS Press and the authors. All rights reserved 2R. Mirizzi, A. Ragone, T. Di Noia, E. Di Sciascio / Lookup, Explore, Discover: how DBpedia can improve your Web search tools take into account the meaning/semantics as- { a hybrid approach to rank resources in a RDF sociated to tags. In this paper we present LED dataset (such as DBpedia): our system ex- (acronym of Lookup Explore Discover), a tool for ploits both the graph nature of a RDF dataset exploratory search exploiting the knowledge en- and text-based techniques used in IR; coded in the Web of data for the tag cloud gener- { a RESTful JSON API that can be accessed ation. Although LED shares with the above men- from any environment that allows HTTP re- tioned tools all the main functionalities such as quest, useful both for building innovative web add/remove tags in order to refine the query, the applications and for comparison with different way it generates the tag cloud is completely dif- algorithms. ferent: The remainder of this paper is structured as fol- { tags are semantically related to the query and lows: in Section 2 we present in detail the LED in- not to the (partial) content of search results; terface. Section 3 is focused on the back-end of the { each generated tag is associated to a RDF re- system. Section 4 shows how LED backend can be source coming from the Web of Data (in our used as a semantic tagging system. In Section 5 we case we mainly rely on DBpedia [2]) with its present the LED RESTful API. A discussion and a associate semantics; comparison of related work is done in Section 6. { tags are semantically related with each other; Conclusion and future work close the paper. Therefore, while for the other tools the main aim is to give to the user a \visual summary" of the 2. LED: Lookup Explore Discover search results, in LED the terms included in the tag cloud are the most relevant w.r.t. the typed In this section we introduce LED, a web-based keyword(s), not the most relevant w.r.t. the re- tool that helps users in query refinement on trieved documents. Also for LED, the main pur- search engines and exploratory knowledge search pose is to suggest keywords similar to the one(s) in DBpedia. The front-end of the system is avail- representing the query in order to allow users able at http://sisinflab.poliba.it/led/ and to refine it and narrow down the search results currently is focused on search in the IT domain. adding/removing new tags from the cloud. In LED In a nutshell, the web-interface is composed by the different dimension of tags in the cloud does three different areas. The first one (marked with not reflect the number of occurrences of the key- (1) in Fig. 1) is mainly a keyword lookup: here the words in the search results but the semantic simi- user types in some characters to obtain an auto- larity of each tags w.r.t. the searched keyword(s). complete list of related keywords. The second area LED wants to favour a serendipitous approach [6], (marked with (2)) in Fig. 1 consists of a tag cloud that is the fortuitous discovery of something which (or a set of tag clouds) populated with automati- is not already included in the search results and cally generated tags that are semantically related that the user did not know to exist, but that could with the keyword entered by the user. Each of be as well of interest for her. them corresponds to a DBpedia resource. In fact, The innovative aspects of LED are: the displayed tag is the value of the rdf:label of the corresponding resource. The user may also { a novel approach to exploratory browsing that search for multiple keywords thanks to a tab-based exploits the Semantic Web: in this first ver- structure of the area devoted to search results. As sion of the tool we focus on DBpedia and shown in Fig. 1, one tab for each searched keyword present an intuitive way to explore this knowl- is open (RDFa and Microformat, respectively) and edge base; one with all the keywords (RDFa Microformat). { a web application that help users to refine The tag cloud allows the user to discover new queries for search engines with the aid of se- knowledge that is hidden in a traditional search mantically generated suggestions; engine. Thanks to LED it is also possible to explore { a meta-search engine and news collector that knowledge, browsing through suggested tags, and integrates results coming from different exter- refine the query by adding more specific keywords. nal sources in a single platform; Finally, the third areas of the interface (marked R. Mirizzi, A. Ragone, T. Di Noia, E. Di Sciascio / Lookup, Explore, Discover: how DBpedia can improve your Web search3 Fig. 1. The LED interface, available online at http://sisinflab.poliba.it/led/. with (3) in Fig. 1) acts as a meta-search engine: [8]. Therefore, using additional external knowledge it displays the results coming from the three most to refine the queries can be a good way to narrow popular search engines (i.e., Google, Yahoo! and down the search results and improve the user ex- Bing), the social network and microblogging ser- perience, maximizing the number of possibly rele- vice Twitter, the news aggregator Google News, vant retrieved objects. In [11] the authors describe Flickr and Google Images, according to the query the main ideas behind a simple lookup search on a formulated in the first section and refined in the traditional search engine and a deeper (semantic- second section. guided) exploratory search: a search engine allows to find what you are looking for and already know, 2.1. Lookup while LED (and more generally exploratory search activities) allows to explore and discover what you The entry point to LED is a text input field. Sim- probably did not know to exist.
Recommended publications
  • A Survey of Orthographic Information in Machine Translation 3
    Machine Translation Journal manuscript No. (will be inserted by the editor) A Survey of Orthographic Information in Machine Translation Bharathi Raja Chakravarthi1 ⋅ Priya Rani 1 ⋅ Mihael Arcan2 ⋅ John P. McCrae 1 the date of receipt and acceptance should be inserted later Abstract Machine translation is one of the applications of natural language process- ing which has been explored in different languages. Recently researchers started pay- ing attention towards machine translation for resource-poor languages and closely related languages. A widespread and underlying problem for these machine transla- tion systems is the variation in orthographic conventions which causes many issues to traditional approaches. Two languages written in two different orthographies are not easily comparable, but orthographic information can also be used to improve the machine translation system. This article offers a survey of research regarding orthog- raphy’s influence on machine translation of under-resourced languages. It introduces under-resourced languages in terms of machine translation and how orthographic in- formation can be utilised to improve machine translation. We describe previous work in this area, discussing what underlying assumptions were made, and showing how orthographic knowledge improves the performance of machine translation of under- resourced languages. We discuss different types of machine translation and demon- strate a recent trend that seeks to link orthographic information with well-established machine translation methods. Considerable attention is given to current efforts of cog- nates information at different levels of machine translation and the lessons that can Bharathi Raja Chakravarthi [email protected] Priya Rani [email protected] Mihael Arcan [email protected] John P.
    [Show full text]
  • Probabilistic Topic Modelling with Semantic Graph
    Probabilistic Topic Modelling with Semantic Graph B Long Chen( ), Joemon M. Jose, Haitao Yu, Fajie Yuan, and Huaizhi Zhang School of Computing Science, University of Glasgow, Sir Alwyns Building, Glasgow, UK [email protected] Abstract. In this paper we propose a novel framework, topic model with semantic graph (TMSG), which couples topic model with the rich knowledge from DBpedia. To begin with, we extract the disambiguated entities from the document collection using a document entity linking system, i.e., DBpedia Spotlight, from which two types of entity graphs are created from DBpedia to capture local and global contextual knowl- edge, respectively. Given the semantic graph representation of the docu- ments, we propagate the inherent topic-document distribution with the disambiguated entities of the semantic graphs. Experiments conducted on two real-world datasets show that TMSG can significantly outperform the state-of-the-art techniques, namely, author-topic Model (ATM) and topic model with biased propagation (TMBP). Keywords: Topic model · Semantic graph · DBpedia 1 Introduction Topic models, such as Probabilistic Latent Semantic Analysis (PLSA) [7]and Latent Dirichlet Analysis (LDA) [2], have been remarkably successful in ana- lyzing textual content. Specifically, each document in a document collection is represented as random mixtures over latent topics, where each topic is character- ized by a distribution over words. Such a paradigm is widely applied in various areas of text mining. In view of the fact that the information used by these mod- els are limited to document collection itself, some recent progress have been made on incorporating external resources, such as time [8], geographic location [12], and authorship [15], into topic models.
    [Show full text]
  • Word Representation Or Word Embedding in Persian Texts Siamak Sarmady Erfan Rahmani
    1 Word representation or word embedding in Persian texts Siamak Sarmady Erfan Rahmani (Abstract) Text processing is one of the sub-branches of natural language processing. Recently, the use of machine learning and neural networks methods has been given greater consideration. For this reason, the representation of words has become very important. This article is about word representation or converting words into vectors in Persian text. In this research GloVe, CBOW and skip-gram methods are updated to produce embedded vectors for Persian words. In order to train a neural networks, Bijankhan corpus, Hamshahri corpus and UPEC corpus have been compound and used. Finally, we have 342,362 words that obtained vectors in all three models for this words. These vectors have many usage for Persian natural language processing. and a bit set in the word index. One-hot vector method 1. INTRODUCTION misses the similarity between strings that are specified by their characters [3]. The Persian language is a Hindi-European language The second method is distributed vectors. The words where the ordering of the words in the sentence is in the same context (neighboring words) may have subject, object, verb (SOV). According to various similar meanings (For example, lift and elevator both structures in Persian language, there are challenges that seem to be accompanied by locks such as up, down, cannot be found in languages such as English [1]. buildings, floors, and stairs). This idea influences the Morphologically the Persian is one of the hardest representation of words using their context. Suppose that languages for the purposes of processing and each word is represented by a v dimensional vector.
    [Show full text]
  • Semantic Memory: a Review of Methods, Models, and Current Challenges
    Psychonomic Bulletin & Review https://doi.org/10.3758/s13423-020-01792-x Semantic memory: A review of methods, models, and current challenges Abhilasha A. Kumar1 # The Psychonomic Society, Inc. 2020 Abstract Adult semantic memory has been traditionally conceptualized as a relatively static memory system that consists of knowledge about the world, concepts, and symbols. Considerable work in the past few decades has challenged this static view of semantic memory, and instead proposed a more fluid and flexible system that is sensitive to context, task demands, and perceptual and sensorimotor information from the environment. This paper (1) reviews traditional and modern computational models of seman- tic memory, within the umbrella of network (free association-based), feature (property generation norms-based), and distribu- tional semantic (natural language corpora-based) models, (2) discusses the contribution of these models to important debates in the literature regarding knowledge representation (localist vs. distributed representations) and learning (error-free/Hebbian learning vs. error-driven/predictive learning), and (3) evaluates how modern computational models (neural network, retrieval- based, and topic models) are revisiting the traditional “static” conceptualization of semantic memory and tackling important challenges in semantic modeling such as addressing temporal, contextual, and attentional influences, as well as incorporating grounding and compositionality into semantic representations. The review also identifies new challenges
    [Show full text]
  • Rdfa in XHTML: Syntax and Processing Rdfa in XHTML: Syntax and Processing
    RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing A collection of attributes and processing rules for extending XHTML to support RDF W3C Recommendation 14 October 2008 This version: http://www.w3.org/TR/2008/REC-rdfa-syntax-20081014 Latest version: http://www.w3.org/TR/rdfa-syntax Previous version: http://www.w3.org/TR/2008/PR-rdfa-syntax-20080904 Diff from previous version: rdfa-syntax-diff.html Editors: Ben Adida, Creative Commons [email protected] Mark Birbeck, webBackplane [email protected] Shane McCarron, Applied Testing and Technology, Inc. [email protected] Steven Pemberton, CWI Please refer to the errata for this document, which may include some normative corrections. This document is also available in these non-normative formats: PostScript version, PDF version, ZIP archive, and Gzip’d TAR archive. The English version of this specification is the only normative version. Non-normative translations may also be available. Copyright © 2007-2008 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply. Abstract The current Web is primarily made up of an enormous number of documents that have been created using HTML. These documents contain significant amounts of structured data, which is largely unavailable to tools and applications. When publishers can express this data more completely, and when tools can read it, a new world of user functionality becomes available, letting users transfer structured data between applications and web sites, and allowing browsing applications to improve the user experience: an event on a web page can be directly imported - 1 - How to Read this Document RDFa in XHTML: Syntax and Processing into a user’s desktop calendar; a license on a document can be detected so that users can be informed of their rights automatically; a photo’s creator, camera setting information, resolution, location and topic can be published as easily as the original photo itself, enabling structured search and sharing.
    [Show full text]
  • Knowledge Graphs on the Web – an Overview Arxiv:2003.00719V3 [Cs
    January 2020 Knowledge Graphs on the Web – an Overview Nicolas HEIST, Sven HERTLING, Daniel RINGLER, and Heiko PAULHEIM Data and Web Science Group, University of Mannheim, Germany Abstract. Knowledge Graphs are an emerging form of knowledge representation. While Google coined the term Knowledge Graph first and promoted it as a means to improve their search results, they are used in many applications today. In a knowl- edge graph, entities in the real world and/or a business domain (e.g., people, places, or events) are represented as nodes, which are connected by edges representing the relations between those entities. While companies such as Google, Microsoft, and Facebook have their own, non-public knowledge graphs, there is also a larger body of publicly available knowledge graphs, such as DBpedia or Wikidata. In this chap- ter, we provide an overview and comparison of those publicly available knowledge graphs, and give insights into their contents, size, coverage, and overlap. Keywords. Knowledge Graph, Linked Data, Semantic Web, Profiling 1. Introduction Knowledge Graphs are increasingly used as means to represent knowledge. Due to their versatile means of representation, they can be used to integrate different heterogeneous data sources, both within as well as across organizations. [8,9] Besides such domain-specific knowledge graphs which are typically developed for specific domains and/or use cases, there are also public, cross-domain knowledge graphs encoding common knowledge, such as DBpedia, Wikidata, or YAGO. [33] Such knowl- edge graphs may be used, e.g., for automatically enriching data with background knowl- arXiv:2003.00719v3 [cs.AI] 12 Mar 2020 edge to be used in knowledge-intensive downstream applications.
    [Show full text]
  • Language Resource Management System for Asian Wordnet Collaboration and Its Web Service Application
    Language Resource Management System for Asian WordNet Collaboration and Its Web Service Application Virach Sornlertlamvanich1,2, Thatsanee Charoenporn1,2, Hitoshi Isahara3 1Thai Computational Linguistics Laboratory, NICT Asia Research Center, Thailand 2National Electronics and Computer Technology Center, Pathumthani, Thailand 3National Institute of Information and Communications Technology, Japan E-mail: [email protected], [email protected], [email protected] Abstract This paper presents the language resource management system for the development and dissemination of Asian WordNet (AWN) and its web service application. We develop the platform to establish a network for the cross language WordNet development. Each node of the network is designed for maintaining the WordNet for a language. Via the table that maps between each language WordNet and the Princeton WordNet (PWN), the Asian WordNet is realized to visualize the cross language WordNet between the Asian languages. We propose a language resource management system, called WordNet Management System (WNMS), as a distributed management system that allows the server to perform the cross language WordNet retrieval, including the fundamental web service applications for editing, visualizing and language processing. The WNMS is implemented on a web service protocol therefore each node can be independently maintained, and the service of each language WordNet can be called directly through the web service API. In case of cross language implementation, the synset ID (or synset offset) defined by PWN is used to determined the linkage between the languages. English PWN. Therefore, the original structural 1. Introduction information is inherited to the target WordNet through its sense translation and sense ID. The AWN finally Language resource becomes crucial in the recent needs for cross cultural information exchange.
    [Show full text]
  • Large Semantic Network Manual Annotation 1 Introduction
    Large Semantic Network Manual Annotation V´aclav Nov´ak Institute of Formal and Applied Linguistics Charles University, Prague [email protected] Abstract This abstract describes a project aiming at manual annotation of the content of natural language utterances in a parallel text corpora. The formalism used in this project is MultiNet – Multilayered Ex- tended Semantic Network. The annotation should be incorporated into Prague Dependency Treebank as a new annotation layer. 1 Introduction A formal specification of the semantic content is the aim of numerous semantic approaches such as TIL [6], DRT [9], MultiNet [4], and others. As far as we can tell, there is no large “real life” text corpora manually annotated with such markup. The projects usually work only with automatically generated annotation, if any [1, 6, 3, 2]. We want to create a parallel Czech-English corpora of texts annotated with the corresponding semantic network. 1.1 Prague Dependency Treebank From the linguistic viewpoint there language resources such as Prague Dependency Treebank (PDT) which contain a deep manual analysis of texts [8]. PDT contains annotations of three layers, namely morpho- logical, analytical (shallow dependency syntax) and tectogrammatical (deep dependency syntax). The units of each annotation level are linked with corresponding units on the preceding level. The morpho- logical units are linked directly with the original text. The theoretical basis of the treebank lies in the Functional Gener- ative Description of language system [7]. PDT 2.0 is based on the long-standing Praguian linguistic tradi- tion, adapted for the current computational-linguistics research needs. The corpus itself is embedded into the latest annotation technology.
    [Show full text]
  • Mapping Between Digital Identity Ontologies Through SISM
    Mapping between Digital Identity Ontologies through SISM Matthew Rowe The OAK Group, Department of Computer Science, University of Sheffield, Regent Court, 211 Portobello Street, Sheffield S1 4DP, UK [email protected] Abstract. Various ontologies are available defining the semantics of dig- ital identity information. Due to the rise in use of lowercase semantics, such ontologies are now used to add metadata to digital identity informa- tion within web pages. However concepts exist in these ontologies which are related and must be mapped together in order to enhance machine- readability of identity information on the web. This paper presents the Social identity Schema Mapping (SISM) vocabulary which contains a set of mappings between related concepts in distinct digital identity ontolo- gies using OWL and SKOS mapping constructs. Key words: Semantic Web, Social Web, SKOS, OWL, FOAF, SIOC, PIMO, NCO, Microformats 1 Introduction The semantic web provides a web of machine-readable data. Ontologies form a vital component of the semantic web by providing conceptualisations of domains of knowledge which can then be used to provide a common understanding of some domain. A basic ontology contains a vocabulary of concepts and definitions of the relationships between those concepts. An agent reading a concept from an ontology can look up the concept and discover its properties and characteristics, therefore interpreting how it fits into that particular domain. Due to the great number of ontologies it is common for related concepts to be defined in separate ontologies, these concepts must be identified and mapped together. Web technologies such as Microformats, eRDF and RDFa have allowed web developers to encode lowercase semantics within XHTML pages.
    [Show full text]
  • Universal Or Variation? Semantic Networks in English and Chinese
    Universal or variation? Semantic networks in English and Chinese Understanding the structures of semantic networks can provide great insights into lexico- semantic knowledge representation. Previous work reveals small-world structure in English, the structure that has the following properties: short average path lengths between words and strong local clustering, with a scale-free distribution in which most nodes have few connections while a small number of nodes have many connections1. However, it is not clear whether such semantic network properties hold across human languages. In this study, we investigate the universal structures and cross-linguistic variations by comparing the semantic networks in English and Chinese. Network description To construct the Chinese and the English semantic networks, we used Chinese Open Wordnet2,3 and English WordNet4. The two wordnets have different word forms in Chinese and English but common word meanings. Word meanings are connected not only to word forms, but also to other word meanings if they form relations such as hypernyms and meronyms (Figure 1). 1. Cross-linguistic comparisons Analysis The large-scale structures of the Chinese and the English networks were measured with two key network metrics, small-worldness5 and scale-free distribution6. Results The two networks have similar size and both exhibit small-worldness (Table 1). However, the small-worldness is much greater in the Chinese network (σ = 213.35) than in the English network (σ = 83.15); this difference is primarily due to the higher average clustering coefficient (ACC) of the Chinese network. The scale-free distributions are similar across the two networks, as indicated by ANCOVA, F (1, 48) = 0.84, p = .37.
    [Show full text]
  • Linked Data Schemata: Fixing Unsound Foundations
    Linked data schemata: fixing unsound foundations. Kevin Feeney, Gavin Mendel Gleason, Rob Brennan Knowledge and Data Engineering Group & ADAPT Centre, School of Computer Science & Statistics, Trinity College Dublin, Ireland Abstract. This paper describes an analysis, and the tools and methods used to produce it, of the practical and logical implications of unifying common linked data vocabularies into a single logical model. In order to support any type of reasoning or even just simple type-checking, the vocabularies that are referenced by linked data statements need to be unified into a complete model wherever they reference or reuse terms that have been defined in other linked data vocabularies. Strong interdependencies between vocabularies are common and a large number of logical and practical problems make this unification inconsistent and messy. However, the situation is far from hopeless. We identify a minimal set of necessary fixes that can be carried out to make a large number of widely-deployed vocabularies mutually compatible, and a set of wider-ranging recommendations for linked data ontology design best practice to help alleviate the problem in future. Finally we make some suggestions for improving OWL’s support for distributed authoring and ontology reuse in the wild. Keywords: Linked Data, Reasoning, Data Quality 1. Introduction One of the central tenets of the Linked Data movement is the reuse of terms from existing well- known vocabularies [Bizer09] when developing new schemata or datasets. The semantic web infrastructure, and the RDF, RDFS and OWL languages, support this with their inherently distributed and modular nature. In practice, through vocabulary reuse, linked data schemata adopt knowledge models that are based on multiple, independently devised ontologies that often exhibit varying definitional semantics [Hogan12].
    [Show full text]
  • Extracting the Features of Modern Web Applications Based on Web Engineering Methods
    (IJACSA) International Journal of Advanced Computer Science and Applications, Vol. 10, No. 2, 2019 Extracting the Features of Modern Web Applications based on Web Engineering Methods 1 Karzan Wakil Dayang N.A. Jawawi2 Research Center, Sulaimani Polytechnic University Software Engineering Department Sulaimani 46001, Kurdistan Region, Iraq Faculty of Computing Software Engineering Department, Faculty of Computing Universiti Teknologi Malaysia 81310 Universiti Teknologi Malaysia, Johor, Malaysia Johor Bahru, Johor, Malaysia Abstract—With the revolution of the information, an applications [7], in this paper imported web 5.0 to web 4.0 advanced version of the web proposed from web 1.0 to web 4.0. because of both generations regarded to Artificial Intelligent In each version, many web applications appeared. In the new (AI). Clusters of webs 3.0 and 4.0 represent the MWAs that versions, modern web applications (MWAs) proposed. These possess great extent of complexities, encompassing Ubiquitous applications have specific features and different features, and Web Applications (UWAs), Rich Internet Applications (RIAs), these features made a new challenge for web engineering Semantic Web Applications (SWAs), and Intelligent Web methods. The problem is that web engineering methods have Applications (IWAs). limitations for MWAs, and the gap is that the developers cannot highlight the new features based on web engineering methods. In Model-Driven Web Engineering (MDWE) is deployed this paper, we extract features of the MWA based on web based on the concept of separation of representation models in engineering methods. We extract web application modules for designing web applications, which is advantageous, showing interaction and structure of their feature based on predominantly, as the platforms and technologies employed in models and elements of web engineering methods.
    [Show full text]