Meta Search Engine with an Intelligent Interface for Information Retrieval on Multiple Domains
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Internet Training: the Basics. INSTITUTION Temple Univ., Philadelphia, PA
DOCUMENT RESUME ED 424 360 CE 077 241 AUTHOR Gallo, Gail; Wechowski, Chester P. TITLE Internet Training: The Basics. INSTITUTION Temple Univ., Philadelphia, PA. Center for Vocational Education Professional Personnel Development. PUB DATE 1998-04-00 NOTE 27p. PUB TYPE Guides Non-Classroom (055) EDRS PRICE MF01/PCO2 Plus Postage. DESCRIPTORS Adult Education; Educational Needs; *Information Sources; *Online Systems; Postsecondary Education; Secondary Education; Teaching Methods; Vocational Education Teachers; *World Wide Web IDENTIFIERS Web Sites ABSTRACT This paper outlines the basic information teachers need to . know to use the World Wide Web for research and communication, using Netscape 3.04. Topics covered include the following: what is the World Wide Web?; what is a browser?; accessing the Web; moving around a web document; the Uniform Resource Locator (URL); Bookmarks; saving and printing a web page; Netscape's options preferences menu; searching the Web; and interesting places (websites of interest to vocational educators) .Two appendixes contain the following: "Infoseek: How Do I Search? 'Ruby Slippers'" and Guidelines for Completing Abstracts. (KC) ******************************************************************************** * Reproductions supplied by EDRS are the best that can be made * * from the original document. * ***********************************************************************i******** t. The Center for Vocational Education \I/ Professional Personnel Development/ \ INTERNET TRAINING: THE BASICS Developed by Gail Gallo Project Director Chester P. Wichowski Temple University April, 1998 ./1.S. DEPARTMENT OFEDUCATION /ffice of Educational Research and Improvement PERMISSION TO REPRODUCE AND E UCATIONAL RESOURCES INFORMATION DISSEMINATE THIS MATERIAL HAS CENTER (ERIC) BEEN GRANTED BY This document has been reproduced as received from the person or organization originating it. 0 Minor changes have been made to improve reproduction quality. -
Market Research SD-5 Gathering Information About Commercial Products and Services
Market Research SD-5 Gathering Information About Commercial Products and Services DEFENSE STANDARDIZATION PROGRA M JANUARY 2008 Contents Foreword 1 The Market Research Other Considerations 32 Background 2 Process 13 Amount of Information Strategic Market Research to Gather 32 What Is Market Research? 2 (Market Surveillance) 14 Procurement Integrity Act 32 Why Do Market Research? 2 Identify the Market or Market Paperwork Reduction Act 33 Segment of Interest 14 When Is Market Research Cost of Market Research 34 Done? 5 Identify Sources of Market Information 16 Who Should Be Involved In Market Research? 7 Collect Relevant Market Other Information Information 17 Technical Specialist 8 Document the Results 18 on Market Research 35 User 9 Logistics Specialist 9 Tactical Market Research Appendix A 36 (Market Investigation) 19 Testing Specialist 9 Types of Information Summarize Strategic Market Available on the Internet Cost Analyst 10 Research 19 Legal Counsel 10 Formulate Requirements 20 Appendix B 39 Contracting Officer 10 Web-Based Information Identify Sources of Sources Information 21 Guiding Principles 11 Collect Product or Service Appendix C 47 Examples of Tactical Start Early 11 Information from Sources 22 Collect Information from Information Define and Document Product or Service Users 26 Requirements 11 Evaluate the Data 27 Refine as You Proceed 12 Document the Results 30 Tailor the Investigation 12 Repeat as Necessary 12 Communicate 12 Involve Users 12 Foreword The Department of Defense (DoD) relies extensively on the commercial market for the products and services it needs, whether those products and services are purely commercial, modified for DoD use from commercial products and services, or designed specifically for DoD. -
Building a Scalable Index and a Web Search Engine for Music on the Internet Using Open Source Software
Department of Information Science and Technology Building a Scalable Index and a Web Search Engine for Music on the Internet using Open Source software André Parreira Ricardo Thesis submitted in partial fulfillment of the requirements for the degree of Master in Computer Science and Business Management Advisor: Professor Carlos Serrão, Assistant Professor, ISCTE-IUL September, 2010 Acknowledgments I should say that I feel grateful for doing a thesis linked to music, an art which I love and esteem so much. Therefore, I would like to take a moment to thank all the persons who made my accomplishment possible and hence this is also part of their deed too. To my family, first for having instigated in me the curiosity to read, to know, to think and go further. And secondly for allowing me to continue my studies, providing the environment and the financial means to make it possible. To my classmate André Guerreiro, I would like to thank the invaluable brainstorming, the patience and the help through our college years. To my friend Isabel Silva, who gave me a precious help in the final revision of this document. Everyone in ADETTI-IUL for the time and the attention they gave me. Especially the people over Caixa Mágica, because I truly value the expertise transmitted, which was useful to my thesis and I am sure will also help me during my professional course. To my teacher and MSc. advisor, Professor Carlos Serrão, for embracing my will to master in this area and for being always available to help me when I needed some advice. -
Awareness and Utilization of Search Engines for Information Retrieval by Students of National Open University of Nigeria in Enugu Study Centre Library
University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Library Philosophy and Practice (e-journal) Libraries at University of Nebraska-Lincoln 2020 Awareness and Utilization of Search Engines for Information Retrieval by Students of National Open University of Nigeria in Enugu Study Centre Library SUNDAY JUDE ONUH National Open University of Nigeria, [email protected] OGOEGBUNAM LOVETH EKWUEME National Open University of Nigeria, [email protected] Follow this and additional works at: https://digitalcommons.unl.edu/libphilprac Part of the Library and Information Science Commons ONUH, SUNDAY JUDE and EKWUEME, OGOEGBUNAM LOVETH, "Awareness and Utilization of Search Engines for Information Retrieval by Students of National Open University of Nigeria in Enugu Study Centre Library" (2020). Library Philosophy and Practice (e-journal). 4650. https://digitalcommons.unl.edu/libphilprac/4650 Awareness and Utilization of Search Engines for Information Retrieval by Students of National Open University of Nigeria in Enugu Study Centre Library By Jude Sunday Onuh Enugu Study Centre Library National Open University of Nigeria [email protected] & Loveth Ogoegbunam Ekwueme Department of Library and Information Science National Open University of Nigeria [email protected] Abstract This study dwelt on awareness and utilization of search engines for information retrieval by students of National Open University of Nigeria (NOUN) Enugu Study centre. Descriptive survey research was adopted for the study. Two research questions were drawn from the two purposes that guided the study. The population consists of 5855 undergraduate students of NOUN Enugu Study Centre. A sample size of 293 students was used as 5% of the entire population. -
A Study on Vertical and Broad-Based Search Engines
International Journal of Latest Trends in Engineering and Technology IJLTET Special Issue- ICRACSC-2016 , pp.087-093 e-ISSN: 2278-621X A STUDY ON VERTICAL AND BROAD-BASED SEARCH ENGINES M.Swathi1 and M.Swetha2 Abstract-Vertical search engines or Domain-specific search engines[1][2] are becoming increasingly popular because they offer increased accuracy and extra features not possible with general, Broad-based search engines or Web-wide search engines. The paper focuses on the survey of domain specific search engine which is becoming more popular as compared to Web- Wide Search Engines as they are difficult to maintain and time consuming .It is also difficult to provide appropriate documents to represent the target data. We also listed various vertical search engines and Broad-based search engines. Index terms: Domain specific search, vertical search engines, broad based search engines. I. INTRODUCTION The Web has become a very rich source of information for almost any field, ranging from music to histories, from sports to movies, from science to culture, and many more. However, it has become increasingly difficult to search for desired information on the Web. Users are facing the problem of information overload , in which a search on a general-purpose search engine such as Google (www.google.com) results in thousands of hits.Because a user cannot specify a search domain (e.g. medicine, music), a search query may bring up Web pages both within and outside the desired domain. Example 1: A user searching for “cancer” may get Web pages related to the disease as well as those related to the Zodiac sign. -
Search Techniques: Web Based Search MODULE - 5B INFORMATION RETRIEVAL SYSTEM
Search Techniques: Web Based Search MODULE - 5B INFORMATION RETRIEVAL SYSTEM 18 Notes SEARCH TECHNIQUES: WEB BASED SEARCH 18.1 INTRODUCTION Internet has become the biggest repository of information in the world. It can be considered as a global library where variety of information in different languages and formats is stored in digital form. The volume of information on web is enormous and it has become near to impossible to estimate its size. Because of its size and storing mechanism, finding relevant and precise information has become a difficult task. For searching information from this vast repository, we use search engines. There are thousands of search engines available on internet. For example, if you visit http://www.thesearchen ginelist.com/, you will find a classified list of search engines. This list is category-wise and includes all-purpose search engines in various fields like accounting, blogs, books, legal, medical, etc. In Lesson 17, you studied basic concepts of search techniques. Here, you will learn various aspects of searching information on web. 18.2 OBJECTIVES After studying this lesson, you will be able to: explain purpose of simple and advanced search techniques; develop search string using Boolean logic on a given topic; illustrate search string with the help of a diagram; give examples of simple search and advanced search on internet; identify various Search Engines, viz. Google, Yahoo, Google Scholar; LIBRARY AND INFORMATION SCIENCE 367 MODULE - 5B Search Techniques: Web Based Search INFORMATION RETRIEVAL SYSTEM identify Search Engines on internet in different vernacular languages; illustrate search in specific categories, viz. maps, images; and modify search strings to get precise results. -
Evaluation of Web-Based Search Engines Using User-Effort Measures
Evaluation of Web-Based Search Engines Using User-Effort Measures Muh-Chyun Tang and Ying Sun 4 Huntington St. School of Information, Communication and Library Studies Rutgers University, New Brunswick, NJ 08901, U.S.A. [email protected] [email protected] Abstract This paper presents a study of the applicability of three user-effort-sensitive evaluation measures —“first 20 full precision,” “search length,” and “rank correlation”—on four Web-based search engines (Google, AltaVista, Excite and Metacrawler). The authors argue that these measures are better alternatives than precision and recall in Web search situations because of their emphasis on the quality of ranking. Eight sets of search topics were collected from four Ph.D. students in four different disciplines (biochemistry, industrial engineering, economics, and urban planning). Each participant was asked to provide two topics along with the corresponding query terms. Their relevance and credibility judgment of the Web pages were then used to compare the performance of the search engines using these three measures. The results show consistency among these three ranking evaluation measures, more so between “first 20 full precision” and search length than between rank correlation and the other two measures. Possible reasons for rank correlation’s disagreement with the other two measures are discussed. Possible future research to improve these measures is also addressed. Introduction The explosive growth of information on the World Wide Web poses a challenge to traditional information retrieval (IR) research. Other than the sheer amount of information, some structural factors make searching for relevant and quality information on the Web a formidable task. -
Open Search Environments: the Free Alternative to Commercial Search Services
Open Search Environments: The Free Alternative to Commercial Search Services. Adrian O’Riordan ABSTRACT Open search systems present a free and less restricted alternative to commercial search services. This paper explores the space of open search technology, looking in particular at lightweight search protocols and the issue of interoperability. A description of current protocols and formats for engineering open search applications is presented. The suitability of these technologies and issues around their adoption and operation are discussed. This open search approach is especially useful in applications involving the harvesting of resources and information integration. Principal among the technological solutions are OpenSearch, SRU, and OAI-PMH. OpenSearch and SRU realize a federated model to enable content providers and search clients communicate. Applications that use OpenSearch and SRU are presented. Connections are made with other pertinent technologies such as open-source search software and linking and syndication protocols. The deployment of these freely licensed open standards in web and digital library applications is now a genuine alternative to commercial and proprietary systems. INTRODUCTION Web search has become a prominent part of the Internet experience for millions of users. Companies such as Google and Microsoft offer comprehensive search services to users free with advertisements and sponsored links, the only reminder that these are commercial enterprises. Businesses and developers on the other hand are restricted in how they can use these search services to add search capabilities to their own websites or for developing applications with a search feature. The closed nature of the leading web search technology places barriers in the way of developers who want to incorporate search functionality into applications. -
Efficient Focused Web Crawling Approach for Search Engine
Ayar Pranav et al, International Journal of Computer Science and Mobile Computing, Vol.4 Issue.5, May- 2015, pg. 545-551 Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320–088X IJCSMC, Vol. 4, Issue. 5, May 2015, pg.545 – 551 RESEARCH ARTICLE Efficient Focused Web Crawling Approach for Search Engine 1 2 Ayar Pranav , Sandip Chauhan Computer & Science Engineering, Kalol Institute of Technology and Research Canter, Kalol, Gujarat, India 1 [email protected]; 2 [email protected] Abstract— a focused crawler traverses the web, selecting out relevant pages to a predefined topic and neglecting those out of concern. Collecting domain specific documents using focused crawlers has been considered one of most important strategies to find relevant information. While surfing the internet, it is difficult to deal with irrelevant pages and to predict which links lead to quality pages. However most focused crawler use local search algorithm to traverse the web space, but they could easily trapped within limited a sub graph of the web that surrounds the starting URLs also there is problem related to relevant pages that are miss when no links from the starting URLs. There is some relevant pages are miss. To address this problem we design a focused crawler where calculating the frequency of the topic keyword also calculate the synonyms and sub synonyms of the keyword. The weight table is constructed according to the user query. To check the similarity of web pages with respect to topic keywords and priority of extracted link is calculated. -
Distributed Indexing/Searching Workshop Agenda, Attendee List, and Position Papers
Distributed Indexing/Searching Workshop Agenda, Attendee List, and Position Papers Held May 28-19, 1996 in Cambridge, Massachusetts Sponsored by the World Wide Web Consortium Workshop co-chairs: Michael Schwartz, @Home Network Mic Bowman, Transarc Corp. This workshop brings together a cross-section of people concerned with distributed indexing and searching, to explore areas of common concern where standards might be defined. The Call For Participation1 suggested particular focus on repository interfaces that support efficient and powerful distributed indexing and searching. There was a great deal of interest in this workshop. Because of our desire to limit attendance to a workable size group while maximizing breadth of attendee backgrounds, we limited attendance to one person per position paper, and furthermore we limited attendance to one person per institution. In some cases, attendees submitted multiple position papers, with the intention of discussing each of their projects or ideas that were relevant to the workshop. We had not anticipated this interpretation of the Call For Participation; our intention was to use the position papers to select participants, not to provide a forum for enumerating ideas and projects. As a compromise, we decided to choose among the submitted papers and allow multiple position papers per person and per institution, but to restrict attendance as noted above. Hence, in the paper list below there are some cases where one author or institution has multiple position papers. 1 http://www.w3.org/pub/WWW/Search/960528/cfp.html 1 Agenda The Distributed Indexing/Searching Workshop will span two days. The first day's goal is to identify areas for potential standardization through several directed discussion sessions. -
Natural Language Processing Technique for Information Extraction and Analysis
International Journal of Research Studies in Computer Science and Engineering (IJRSCSE) Volume 2, Issue 8, August 2015, PP 32-40 ISSN 2349-4840 (Print) & ISSN 2349-4859 (Online) www.arcjournals.org Natural Language Processing Technique for Information Extraction and Analysis T. Sri Sravya1, T. Sudha2, M. Soumya Harika3 1 M.Tech (C.S.E) Sri Padmavati Mahila Visvavidyalayam (Women’s University), School of Engineering and Technology, Tirupati. [email protected] 2 Head (I/C) of C.S.E & IT Sri Padmavati Mahila Visvavidyalayam (Women’s University), School of Engineering and Technology, Tirupati. [email protected] 3 M. Tech C.S.E, Assistant Professor, Sri Padmavati Mahila Visvavidyalayam (Women’s University), School of Engineering and Technology, Tirupati. [email protected] Abstract: In the current internet era, there are a large number of systems and sensors which generate data continuously and inform users about their status and the status of devices and surroundings they monitor. Examples include web cameras at traffic intersections, key government installations etc., seismic activity measurement sensors, tsunami early warning systems and many others. A Natural Language Processing based activity, the current project is aimed at extracting entities from data collected from various sources such as social media, internet news articles and other websites and integrating this data into contextual information, providing visualization of this data on a map and further performing co-reference analysis to establish linkage amongst the entities. Keywords: Apache Nutch, Solr, crawling, indexing 1. INTRODUCTION In today’s harsh global business arena, the pace of events has increased rapidly, with technological innovations occurring at ever-increasing speed and considerably shorter life cycles. -
Document Selection in a Distributed Search Engine Architecture Ibrahim
Document Selection in a Distributed Search Engine Architecture 1Ibrahim AlShourbaji, 2Samaher Al-Janabi and 3Ahmed Patel 1Computer Network Department, Computer Science and Information System College, Jazan University, Jazan 82822-6649, Saudi Arabia 2Department of Information Networks, Faculty of Information Technology,University of Babylon, Babylon, Hilla 00964, Iraq 3Visiting Professor School of Computing and Information Systems, Faculty of Science, Engineering and Computing, Kingston University, Kingston upon Thames KT1 2EE, United Kingdom Abstract Distributed Search Engine Architecture (DSEA) hosts numerous independent topic-specific search engines and selects a subset of the databases to search within the architecture. The objective of this approach is to reduce the amount of space needed to perform a search by querying only a subset of the total data available. In order to manipulate data across many databases, it is most efficient to identify a smaller subset of databases that would be most likely to return the data of specific interest that can then be examined in greater detail. The selection index has been most commonly used as a method for choosing the most applicable databases as it captures broad information about each database and its indexed documents. Employing this type of database allows the researcher to find information more quickly, not only with less cost, but it also minimizes the potential for biases. This paper investigates the effectiveness of different databases selected within the framework and scope of the distributed search engine architecture. The purpose of the study is to improve the quality of distributed information retrieval. Keywords: web search, distributed search engine, document selection, information retrieval, Collection Retrival Inference network 1.