Searching Web Based Medical Information

Total Page:16

File Type:pdf, Size:1020Kb

Searching Web Based Medical Information Special Review Article Searching web based medical information Ibrahim Mansoor, MBBS. ABSTRACT Searching for references is part of everyday life in medicine. Since the arrival of the Internet, it has provided great promise for clinicians because of it’s ability to access and consolidate large amounts of knowledge and information quickly and easily. But because of the overload of the huge amounts of information, searching for particular information has now become a tedious task, which is a lso very time-consuming and frustrating. This article describes effective ways, tips, tools, detailed search techniques and strategies for searching medical information. It also lists some useful resource and database sites that can provide a lot of help for seeking accurate information. Keywords: Internet, web-based medical information, search engines. Saudi Med J 2001; Vol. 22 (9): 749-756 he Internet is a massive collection of computer do you find what you are looking for? This paper T networks that connect millions of computers, deals with this problem of finding particular medical people, software programs, databases and files. information on the Internet. The paper presents a Though it is comprised of various sections from detailed discussion of different types of search tools, newsgroups through email services, by far, the most with the main features of these major tools, there popular and fastest growing is the World Wide Web strategies, techniques and tips for the users on how to (WWW or just "Web"), cited as most important by interact with these tools, available on the Internet. two-thirds of the users, followed by electronic mail.1,2 Search tools on the web. Finding information on The Web has accelerated the growth of the Internet the Web may be difficult, but it's not impossible. As by giving it an easy to use, point and click, graphical an alternate to a central catalogue the Web offers a interface and it contains a universe of network choice of dozens of different Search Tools, each with accessible information. In February 1999, the Web its own database, command language, search contained approximately 800 million pages of capabilities, and method of displaying results. The information, increased up from 320 million in Search Tools, also known as Search Services, find December 1997.3,4 Adding to this, the fact that the documents matching your interests. Each search tool Web lacks the bibliographic controls standards. operates on its database of Uniform Source Locator There is no equivalent to the ISBN to uniquely (URLs), texts and descriptions that point to the actual identify a document; no standard system of documents on the Web.2,7 An important point to cataloguing or classification; and there is no central clarify is that whenever you search with the help of a catalogue that point to all of the Web's holdings. In search tool, you actually view data extracted from the fact, many, if not most, Web documents lack even database of this search tool and not from the whole the name of the author and the date of publication.5,6 Web. Since none of these search tool databases Now with 800 million publicly available documents include the whole Web, you get different results from — an amount that is doubling in size every 6 to 12 different search tools. All search tools provide the months, the Internet has become a vast, global search results as list of Web documents with storehouse of information. The only problem is: how hypertext links; which when clicked takes us away to From the Department of Histopathology, King AbdulAziz University Hospital, Jeddah, Kingdom of Saudi Arabia. Address correspondence and reprint request to: Dr. Ibrahim Mansoor, c/o Mansoor Ali, PO Box 1432, Jeddah 21431, Kingdom of Saudi Arabia. Fax. +966 (02) 661 3164. Email: [email protected] 749 Searching web-based medical information ... Mansoor that particular Web document from the search tool. most effective for finding general information on The search tools on the Web fall into 2 main popular or scholarly subjects. If you are looking for categories: Subject Directories, which rely heavily on something specific, use a search engine.26 It is also the human element as part of their indexing strategy, important to remember that no one has categorized and Search Engines, which keep human/data the entire web. There are millions of web pages on interaction to a bare minimum. Both use software the Web and it is simply impossible to organize robots called "Spiders" that crawl the Web, everything, especially at the rate at which the Web is newsgroups, and gopher, file transfer protocol (FTP) expanding. Examples of subject directories include: and wide area information servers (WAIS) sites, About.com [http://www. About.com], Look Smart extracting URLs (addresses) and keywords to add to [http://www. looksmart.com], WebCrawler [http:// the search tool's database. Both of these search tools www.webcrawler.com], Yahoo [http:// have benefits and drawbacks, depending on what you www.yahoo.com]. are willing to sacrifice.2,7-9 New MetaCrawlers have Search engines.7-26 The Search Engines are very now emerged, as the best way for querying multiple different from directories. While humans organize engines at once. They do not maintain their one and catalog subject directories, search engines rely on database; instead, they act as middle agents and pass computer programs called spiders or robots to crawl on your query to many major search engines.7-26 the Web from one link to another, and log the words Subject directories. The subject directories are on each page into its database. When any keyword hierarchically organized indexes of subject categories related to a topic is typed into a search box, the that allow the Web searcher to browse through lists search engine scans its database and returns a file of Web sites by subject in search of relevant with links to websites containing the word or words information. They are often called subject "trees" specified. It is important to note that when you are because they start with a few main categories and using a search engine you are not searching the then branch out into subcategories, topics, and Internet "live", as it exists at this very moment. subtopics. For example, if you are looking for any Rather, you are searching a fixed database that has medical information related to "Saudi Arabia" at been compiled some time previous to your search.18,26 Yahoo, select "regional" at the top level, then While all search engines are intended to perform the "countries" at the next level, than "Saudi Arabia" at same task, each perform it in a different way, which the 3rd level, and "medical information" at the 4th leads to sometimes amazingly different results. level. This hierarchy could be represented as: Factors that influence results include the size of the regional / countries / Saudi Arabia / medical database, the frequency of updating, and the search information.27 The subject directories, simply called capabilities. Search engines also differ in their search directories, are compiled and maintained by humans, speed, the design of the search interface, the way in and therefore you can often find it a good starting which they display results, and the amount of help point for your topic. They are also useful for finding they offer. In most cases, search engines are best information on a topic when you don't have a precise used to locate a specific piece of information, such as idea of what you need. Furthermore, because their a known document, an image, or a computer maintenance includes human intervention, they program, rather than a general subject. Examples of greatly reduce the probability of retrieving results out search engines include:26 MSN Search [http:// of context. As the subject directories are categorized, search.msn.com/], AltaVista [http://www.altavista. they give links to the top levels or main pages of a com/], Ask Jeeves [http://www.askjeeves.com/], Web site rather than to individual pages and they Excite [http://www.excite.com/], Fast Search cover only a small fraction of the pages available on [http://www.fast.no/sitesearch.html], Google the Web. They are best suited for searching [http://www.google.com/], HotBot [http://www. information about a general subject for example, hotbot.com/], Infoseek [http://www.infoseek.com/], information required about medical fields such as Lycos, [http://www.lycos.com/], Northern Light pathology, ophthalmology etc or about main organs [http://www.northernlight.com/]. like breast, kidney etc, rather than for a deeper and Meta search engines. The growth in the number more specific piece of information. Many large of search engines has led to the creation of "meta" directories also provide an option to search by a search engines, often referred to as multi-threaded or keyword, which takes us directly to a specific topic multi-search engines. These search engines allow a and usually eliminates the need to work through user to search the databases of several major search numerous levels of topics and subtopics. The engines in parallel (at the same time). They do not directories rely on the judgment and expertise of the crawl the Web or maintain a database of Web pages. people compiling them and here it is obvious that Instead, they act as a middle agent, passing on the subject experts would have different and more query to the major engines like AltaVista, Excite, professional selection criteria than a hobbyist or a Lycos, WebCrawler, Yahoo and HotBot, and then casual user.26 Because directories cover only a small returning the results.23-28 The meta-search engines fraction of the pages available on the Web they are work best with simple searches.
Recommended publications
  • Building a Scalable Index and a Web Search Engine for Music on the Internet Using Open Source Software
    Department of Information Science and Technology Building a Scalable Index and a Web Search Engine for Music on the Internet using Open Source software André Parreira Ricardo Thesis submitted in partial fulfillment of the requirements for the degree of Master in Computer Science and Business Management Advisor: Professor Carlos Serrão, Assistant Professor, ISCTE-IUL September, 2010 Acknowledgments I should say that I feel grateful for doing a thesis linked to music, an art which I love and esteem so much. Therefore, I would like to take a moment to thank all the persons who made my accomplishment possible and hence this is also part of their deed too. To my family, first for having instigated in me the curiosity to read, to know, to think and go further. And secondly for allowing me to continue my studies, providing the environment and the financial means to make it possible. To my classmate André Guerreiro, I would like to thank the invaluable brainstorming, the patience and the help through our college years. To my friend Isabel Silva, who gave me a precious help in the final revision of this document. Everyone in ADETTI-IUL for the time and the attention they gave me. Especially the people over Caixa Mágica, because I truly value the expertise transmitted, which was useful to my thesis and I am sure will also help me during my professional course. To my teacher and MSc. advisor, Professor Carlos Serrão, for embracing my will to master in this area and for being always available to help me when I needed some advice.
    [Show full text]
  • Open Search Environments: the Free Alternative to Commercial Search Services
    Open Search Environments: The Free Alternative to Commercial Search Services. Adrian O’Riordan ABSTRACT Open search systems present a free and less restricted alternative to commercial search services. This paper explores the space of open search technology, looking in particular at lightweight search protocols and the issue of interoperability. A description of current protocols and formats for engineering open search applications is presented. The suitability of these technologies and issues around their adoption and operation are discussed. This open search approach is especially useful in applications involving the harvesting of resources and information integration. Principal among the technological solutions are OpenSearch, SRU, and OAI-PMH. OpenSearch and SRU realize a federated model to enable content providers and search clients communicate. Applications that use OpenSearch and SRU are presented. Connections are made with other pertinent technologies such as open-source search software and linking and syndication protocols. The deployment of these freely licensed open standards in web and digital library applications is now a genuine alternative to commercial and proprietary systems. INTRODUCTION Web search has become a prominent part of the Internet experience for millions of users. Companies such as Google and Microsoft offer comprehensive search services to users free with advertisements and sponsored links, the only reminder that these are commercial enterprises. Businesses and developers on the other hand are restricted in how they can use these search services to add search capabilities to their own websites or for developing applications with a search feature. The closed nature of the leading web search technology places barriers in the way of developers who want to incorporate search functionality into applications.
    [Show full text]
  • Informa2on Retrieval
    Introducon to Informaon Retrieval Introducon to Informaon Retrieval Brief (non‐technical) history . Early keyword‐based engines ca. 1995‐1997 Introducon to . Altavista, Excite, Infoseek, Inktomi, Lycos Informaon Retrieval . Paid search ranking: Goto (morphed into Overture.com → Yahoo!) CS276 . Your search ranking depended on how much you Informaon Retrieval and Web Search paid Pandu Nayak and Prabhakar Raghavan . Aucon for keywords: casino was expensive! Lecture 15: Web search basics 2 Introducon to Informaon Retrieval Introducon to Informaon Retrieval Brief (non‐technical) history . 1998+: Link‐based ranking pioneered by Google . Blew away all early engines save Inktomi . Great user experience in search of a business model . Meanwhile Goto/Overture’s annual revenues were nearing $1 billion Paid . Result: Google added paid search “ads” to the side, Search Ads independent of search results . Yahoo followed suit, acquiring Overture (for paid placement) and Inktomi (for search) . 2005+: Google gains search share, dominang in Europe and very strong in North America . 2009: Yahoo! and Microso propose combined paid search offering Algorithmic results. 3 4 Introducon to Informaon Retrieval Sec. 19.4.1 Introducon to Informaon Retrieval Sec. 19.4.1 Web search basics User Needs . Need [Brod02, RL04] User . Informaonal – want to learn about something (~40% / 65%) . Navigaonal – want to go to that page (~25% / 15%) Web spider . Transaconal – want to do something (web‐mediated) (~35% / 20%) . Access a service . Downloads Search . Shop Indexer . Gray areas . Find a good hub . Exploratory search “see what’s there” The Web Indexes Ad indexes5 6 1 Introducon to Informaon Retrieval Introducon to Informaon Retrieval How far do people look for results? Users’ empirical evaluaon of results .
    [Show full text]
  • Efficient Focused Web Crawling Approach for Search Engine
    Ayar Pranav et al, International Journal of Computer Science and Mobile Computing, Vol.4 Issue.5, May- 2015, pg. 545-551 Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320–088X IJCSMC, Vol. 4, Issue. 5, May 2015, pg.545 – 551 RESEARCH ARTICLE Efficient Focused Web Crawling Approach for Search Engine 1 2 Ayar Pranav , Sandip Chauhan Computer & Science Engineering, Kalol Institute of Technology and Research Canter, Kalol, Gujarat, India 1 [email protected]; 2 [email protected] Abstract— a focused crawler traverses the web, selecting out relevant pages to a predefined topic and neglecting those out of concern. Collecting domain specific documents using focused crawlers has been considered one of most important strategies to find relevant information. While surfing the internet, it is difficult to deal with irrelevant pages and to predict which links lead to quality pages. However most focused crawler use local search algorithm to traverse the web space, but they could easily trapped within limited a sub graph of the web that surrounds the starting URLs also there is problem related to relevant pages that are miss when no links from the starting URLs. There is some relevant pages are miss. To address this problem we design a focused crawler where calculating the frequency of the topic keyword also calculate the synonyms and sub synonyms of the keyword. The weight table is constructed according to the user query. To check the similarity of web pages with respect to topic keywords and priority of extracted link is calculated.
    [Show full text]
  • Distributed Indexing/Searching Workshop Agenda, Attendee List, and Position Papers
    Distributed Indexing/Searching Workshop Agenda, Attendee List, and Position Papers Held May 28-19, 1996 in Cambridge, Massachusetts Sponsored by the World Wide Web Consortium Workshop co-chairs: Michael Schwartz, @Home Network Mic Bowman, Transarc Corp. This workshop brings together a cross-section of people concerned with distributed indexing and searching, to explore areas of common concern where standards might be defined. The Call For Participation1 suggested particular focus on repository interfaces that support efficient and powerful distributed indexing and searching. There was a great deal of interest in this workshop. Because of our desire to limit attendance to a workable size group while maximizing breadth of attendee backgrounds, we limited attendance to one person per position paper, and furthermore we limited attendance to one person per institution. In some cases, attendees submitted multiple position papers, with the intention of discussing each of their projects or ideas that were relevant to the workshop. We had not anticipated this interpretation of the Call For Participation; our intention was to use the position papers to select participants, not to provide a forum for enumerating ideas and projects. As a compromise, we decided to choose among the submitted papers and allow multiple position papers per person and per institution, but to restrict attendance as noted above. Hence, in the paper list below there are some cases where one author or institution has multiple position papers. 1 http://www.w3.org/pub/WWW/Search/960528/cfp.html 1 Agenda The Distributed Indexing/Searching Workshop will span two days. The first day's goal is to identify areas for potential standardization through several directed discussion sessions.
    [Show full text]
  • Natural Language Processing Technique for Information Extraction and Analysis
    International Journal of Research Studies in Computer Science and Engineering (IJRSCSE) Volume 2, Issue 8, August 2015, PP 32-40 ISSN 2349-4840 (Print) & ISSN 2349-4859 (Online) www.arcjournals.org Natural Language Processing Technique for Information Extraction and Analysis T. Sri Sravya1, T. Sudha2, M. Soumya Harika3 1 M.Tech (C.S.E) Sri Padmavati Mahila Visvavidyalayam (Women’s University), School of Engineering and Technology, Tirupati. [email protected] 2 Head (I/C) of C.S.E & IT Sri Padmavati Mahila Visvavidyalayam (Women’s University), School of Engineering and Technology, Tirupati. [email protected] 3 M. Tech C.S.E, Assistant Professor, Sri Padmavati Mahila Visvavidyalayam (Women’s University), School of Engineering and Technology, Tirupati. [email protected] Abstract: In the current internet era, there are a large number of systems and sensors which generate data continuously and inform users about their status and the status of devices and surroundings they monitor. Examples include web cameras at traffic intersections, key government installations etc., seismic activity measurement sensors, tsunami early warning systems and many others. A Natural Language Processing based activity, the current project is aimed at extracting entities from data collected from various sources such as social media, internet news articles and other websites and integrating this data into contextual information, providing visualization of this data on a map and further performing co-reference analysis to establish linkage amongst the entities. Keywords: Apache Nutch, Solr, crawling, indexing 1. INTRODUCTION In today’s harsh global business arena, the pace of events has increased rapidly, with technological innovations occurring at ever-increasing speed and considerably shorter life cycles.
    [Show full text]
  • A Study on the Web Portal Industry
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by KDI School Archives A STUDY ON THE WEB PORTAL INDUSTRY: By Chan-Soo Park THESIS Submitted to School of Public Policy and Management, KDI In partial fulfillment of the requirements for the degree of MASTER OF STRATEGY & GLOBAL MANAGEMENT Department of Strategy & International Management 2000 A STUDY ON THE WEB PORTAL INDUSTRY: By Chan-Soo Park THESIS Submitted to School of Public Policy and Management, KDI In partial fulfillment of the requirements for the degree of MASTER OF STRATEGY & GLOBAL MANAGEMENT Department of Strategy & International Management 2000 Professor Seung-Joo Lee 1 ABSTRACT A STUDY ON THE WEB PORTAL INDUSTRY By Chan –Soo Park A portal is a site on the Internet that provides a one-stop experience for Internet users, allowing them to check e-mail, search the Web, and get personalized news and stock quotes. Since 1998, the “portal” has been considered the most successful Internet business model. Portal sites have characteristics such as community facilitated by services, revenue that relies on advertising, a highly competitive market, heavy traffic, and an uncertain business. The world wide portal industry is like a battle zone for America’s top three, broad-based and vertical portals. The Web portal industry is dominated by the “top three portals,” which are AOL, Yahoo and MSN, and “vertical portals” such as Go Network and NBC. The broad-based portals --Lycos, Excite@home, AltaVista and Infoseek—do not lag far behind as major competitors. Many challenges face the three key players and broad-based portals.
    [Show full text]
  • INSIDE MICROSOFT (Part 2) 2/7/04 2:30 PM
    07/15/96 INSIDE MICROSOFT (Part 2) 2/7/04 2:30 PM INSIDE MICROSOFT (Part 2) The untold story of how the Internet forced Bill Gates to reverse course (Continued from Part 1) BABY STEPS. In one breakout group, Allard tangled with Russell Siegelman, who was heading Marvel, the code name for what's now the Microsoft Network online service. Allard argued that instead of being proprietary, Marvel should be based on Web standards. Siegelman held his ground--and won. It was a decision that would later cost millions to reverse. Still, Net progress was made: TCP/IP would be integrated into Win95 and Windows NT, the version of Windows that runs network-server computers. The sales team was told to use the Web to dispense marketing information. The applications group agreed to give Word, the word-processing program, the ability to create Web pages. Next, Gates jumped deeper into the process by devoting much of his April Think Week--a semiannual retreat--to the Internet. His Apr. 16 memo, ``Internet Strategy and Technical Goals,'' contained the first signs of a growing corporate commitment. ``We want to and will invest resources to be a leader in Internet support,'' wrote Gates. It was a first step, albeit a measured one. ``I don't think he knew how much to bet yet,'' says Allard. But board member David F. Marquardt did: He recalls that he was ``amazed'' that Microsoft was putting so little into the Net. ``They weren't in Silicon Valley. When you're here, you feel it all around you,'' says Marquardt, a general partner at Technology Venture Investors in Menlo Park, Calif.
    [Show full text]
  • An Ontology-Based Web Crawling Approach for the Retrieval of Materials in the Educational Domain
    An Ontology-based Web Crawling Approach for the Retrieval of Materials in the Educational Domain Mohammed Ibrahim1 a and Yanyan Yang2 b 1School of Engineering, University of Portsmouth, Anglesea Road, PO1 3DJ, Portsmouth, United Kingdom 2School of Computing, University of Portsmouth, Anglesea Road, PO1 3DJ, Portsmouth, United Kingdom Keywords: Web Crawling, Ontology, Education Domain. Abstract: As the web continues to be a huge source of information for various domains, the information available is rapidly increasing. Most of this information is stored in unstructured databases and therefore searching for relevant information becomes a complex task and the search for pertinent information within a specific domain is time-consuming and, in all probability, results in irrelevant information being retrieved. Crawling and downloading pages that are related to the user’s enquiries alone is a tedious activity. In particular, crawlers focus on converting unstructured data and sorting this into a structured database. In this paper, among others kind of crawling, we focus on those techniques that extract the content of a web page based on the relations of ontology concepts. Ontology is a promising technique by which to access and crawl only related data within specific web pages or a domain. The methodology proposed is a Web Crawler approach based on Ontology (WCO) which defines several relevance computation strategies with increased efficiency thereby reducing the number of extracted items in addition to the crawling time. It seeks to select and search out web pages in the education domain that matches the user’s requirements. In WCO, data is structured based on the hierarchical relationship, the concepts which are adapted in the ontology domain.
    [Show full text]
  • A Hadoop Based Platform for Natural Language Processing of Web Pages and Documents
    Journal of Visual Languages and Computing 31 (2015) 130–138 Contents lists available at ScienceDirect Journal of Visual Languages and Computing journal homepage: www.elsevier.com/locate/jvlc A hadoop based platform for natural language processing of web pages and documents Paolo Nesi n,1, Gianni Pantaleo 1, Gianmarco Sanesi 1 Distributed Systems and Internet Technology Lab, DISIT Lab, Department of Information Engineering (DINFO), University of Florence, Firenze, Italy article info abstract Available online 30 October 2015 The rapid and extensive pervasion of information through the web has enhanced the Keywords: diffusion of a huge amount of unstructured natural language textual resources. A great Natural language processing interest has arisen in the last decade for discovering, accessing and sharing such a vast Hadoop source of knowledge. For this reason, processing very large data volumes in a reasonable Part-of-speech tagging time frame is becoming a major challenge and a crucial requirement for many commercial Text parsing and research fields. Distributed systems, computer clusters and parallel computing Web crawling paradigms have been increasingly applied in the recent years, since they introduced Big Data Mining significant improvements for computing performance in data-intensive contexts, such as Parallel computing Big Data mining and analysis. Natural Language Processing, and particularly the tasks of Distributed systems text annotation and key feature extraction, is an application area with high computational requirements; therefore, these tasks can significantly benefit of parallel architectures. This paper presents a distributed framework for crawling web documents and running Natural Language Processing tasks in a parallel fashion. The system is based on the Apache Hadoop ecosystem and its parallel programming paradigm, called MapReduce.
    [Show full text]
  • Curlcrawler:A Framework of Curlcrawler and Cached Database for Crawling the Web with Thumb
    Ela Kumaret al, / (IJCSIT) International Journal of Computer Science and Information Technologies, Vol. 2 (4) , 2011, 1700-1705 CurlCrawler:A framework of CurlCrawler and Cached Database for crawling the web with thumb. Dr Ela Kumar#,Ashok Kumar$ # School of Information and Communication Technology, Gautam Buddha University, Greater Noida $ AIM & ACT.Banasthali University,Banasthali(Rajasthan.) Abstract-WWW is a vast source of information and search engines about basic modules of search engine functioning and these are the meshwork to navigate the web for several purposes. It is essential modules are (see Fig.1)[9,10]. problematic to identify and ping with graphical frame of mind for the desired information amongst the large set of web pages resulted by the search engine. With further increase in the size of the Internet, the problem grows exponentially. This paper is an endeavor to develop and implement an efficient framework with multi search agents to make search engines inapt to chaffing using curl features of the programming, called CurlCrawler. This work is an implementation experience for use of graphical perception to upright search on the web. Main features of this framework are curl programming, personalization of information, caching and graphical perception. Keywords and Phrases-Indexing Agent, Filtering Agent, Interacting Agent, Seolinktool, Thumb, Whois, CachedDatabase, IECapture, Searchcon, Main_spider, apnasearchdb. 1. INTRODUCTION Information is a vital role playing versatile thing from availability at church level to web through trends of books. WWW is now the prominent and up-to-date huge repository of information available to everyone, everywhere and every time. Moreover information on web is navigated using search Fig.1: Components of Information Retrieval System engines like AltaVista, WebCrawler, Hot Boat etc[1].Amount of information on the web is growing at very fast and owing to this search engine’s optimized functioning is always a thrust Store: It stores the crawled pages.
    [Show full text]
  • Meta Search Engine with an Intelligent Interface for Information Retrieval on Multiple Domains
    International Journal of Computer Science, Engineering and Information Technology (IJCSEIT), Vol.1, No.4, October 2011 META SEARCH ENGINE WITH AN INTELLIGENT INTERFACE FOR INFORMATION RETRIEVAL ON MULTIPLE DOMAINS D.Minnie1, S.Srinivasan2 1Department of Computer Science, Madras Christian College, Chennai, India [email protected] 2Department of Computer Science and Engineering, Anna University of Technology Madurai, Madurai, India [email protected] ABSTRACT This paper analyses the features of Web Search Engines, Vertical Search Engines, Meta Search Engines, and proposes a Meta Search Engine for searching and retrieving documents on Multiple Domains in the World Wide Web (WWW). A web search engine searches for information in WWW. A Vertical Search provides the user with results for queries on that domain. Meta Search Engines send the user’s search queries to various search engines and combine the search results. This paper introduces intelligent user interfaces for selecting domain, category and search engines for the proposed Multi-Domain Meta Search Engine. An intelligent User Interface is also designed to get the user query and to send it to appropriate search engines. Few algorithms are designed to combine results from various search engines and also to display the results. KEYWORDS Information Retrieval, Web Search Engine, Vertical Search Engine, Meta Search Engine. 1. INTRODUCTION AND RELATED WORK WWW is a huge repository of information. The complexity of accessing the web data has increased tremendously over the years. There is a need for efficient searching techniques to extract appropriate information from the web, as the users require correct and complex information from the web. A Web Search Engine is a search engine designed to search WWW for information about given search query and returns links to various documents in which the search query’s key words are found.
    [Show full text]