Web Searching: a Quality Measurement Perspective to Appear In: Zimmer, M.; Spink, A

Total Page:16

File Type:pdf, Size:1020Kb

Web Searching: a Quality Measurement Perspective to Appear In: Zimmer, M.; Spink, A Web Searching: A Quality Measurement Perspective to appear in: Zimmer, M.; Spink, A. (eds.): Web Search: Interdisciplinary perspec- tives. Dordrecht: Spinger, 2007. Dirk Lewandowski Hamburg University of Applied Sciences, Department Information, Ber- liner Tor 5, D – 20099 Hamburg, Germany. E-Mail: [email protected] Nadine Höchstötter Institute for Decision Theory and Management Science, Universitaet Karlsruhe (TH), Kaiserstrasse 12, D - 76128 Karlsruhe, Germany. E-mail: [email protected] Abstract The purpose of this paper is to describe various quality measures for search engines and to ask whether these are suitable. We especially focus on user needs and their use of web search engines. The paper presents an extensive literature review and a first quality measurement model, as well. Findings include that search engine quality can not be measured by just re- trieval effectiveness (the quality of the results), but should also consider index quality, the quality of the search features and search engine usability. For each of these sections, empirical results from studies conducted in the past, as well as from our own research are presented. These results have implications for the evaluation of search engines and for the development of better search systems that give the user the best possible search experi- ence. Introduction Web search engines have become important for information seeking in many different contexts (e.g., personal, business, and scientific). Research questions not answered satisfactorily are, as of now, how well these en- gines perform regarding user expectations and what measures should be used to get an overall picture of search engine quality. It is well known that search engine quality in its entirety cannot be measured with the use of traditional retrieval measures. But the development of new, search engine specific measures, as proposed in Vaughan (2004) are not sufficient, ei- ther. Search engine quality must be defined more extensively and integrate factors beyond retrieval performance such as index quality and the quality of the search features. One aspect neglected is the user himself. But to discuss and judge the quality of search engines, it is important to focus on the user of such sys- tems, too. A better performance of ranking algorithms or providing addi- tional services do not always lead to users’ satisfaction and to better search results. We will focus on the search engine user behaviour to derive strate- gies to measure search engine quality. Additionally, quality assurance is an important aspect to improve cus- tomer satisfaction and loyalty. This is fundamental to protect market shares and revenues from adverts. Furthermore, quality measurement helps to identify potential improvements of search engines. We are sure that only an integrated approach to quality measurement can lead to results usable for the development of better search engines. As with information retrieval, in general, we find a paradigm shift from the more technical (document-oriented) perspective to the user-oriented per- spective (cf. Ingwersen & Järvelin, 2005). Our goal in this chapter is to de- fine the scope of our perspective in comparison to other approaches and to give a literature overview of quality measurements for search engines. We will also focus on each individual factor stated in studies dealing with user interaction with search engines and user expectations to search engines. The integrated approach of user and technical aspects shows that there are many possibilities but they are not widely adopted yet. Our chapter first gives an overview of studies conducted to derive qual- ity measures and to present the state of the art. The other focus in this sec- tion lies on user surveys and analyses to give an anticipation of what users really do by placing search queries. In section 3 we give a general con- spectus of parameters we deduced from our literature research and explain them shortly. In section 4 we show empirical results that reflect the current quality standard by our individual measures of search engines. In the last 2 section we summarize our findings and give potential strategies to improve search engines. Many of the empirical findings stem from our own research conducted over the past years. Our integrated view on search engine quality meas- urement is reflected by the different research areas of the authors. Related studies In this section, we will discuss studies dealing with search engines in the given context. The two areas relevant for extensive search engine quality measurement are the concept of information quality in general and its transfer to search engines as a technical background, and user studies to see what happens at the front-end. Each will be discussed under a separate heading. Search engine quality Referring to information quality, one usually appraises information on the basis of a single document or a set of documents. Two perspectives have to be differentiated: Firstly, information quality in the production of a database which means, how documents or sources have to be appropriately selected and secondly, information quality of the results retrieved by a cer- tain IR system. While the latter can be easily applied to Web search engines, the assur- ance of the quality of databases is more difficult. The approach of the ma- jor search engines is to index not only a part of the Web, but as much as possible (or as much as reasonable under economic aspects). Only certain fractions of the Web (such as Spam sites) should be willingly omitted from the database. While in the production of databases the process of selecting documents (or sources of documents) can be seen as an important quality aspect, in the context of search engines, this process is reassigned to the ranking process. Therefore, classic judgements for the selection of docu- ments from a library context do not fit to search engines. Only specialized search engines rely on a selection of quality sources (Web sites or servers) for building their indices. An important point is that quality measurement of search results give only limited insight into the reliability and correctness of the information presented in the document. Popular examples are documents from Wikipe- dia, which are often highly ranked by search engines. But there seems not to be an agreement of experts whether Wikipedia content is trustworthy or 3 not. For a normal user, there is only a limited chance of scrutinising these documents. In this context, perceived information quality is more a matter of trust. Within the wider context of search engine evaluation, it is possible to build models completely based on trust (Wang, Xie, & Goh, 1999), as explained later on. When discussing quality of search results, one should also keep in mind how search engines determine relevance. They mainly focus on popularity (or authority) rather than on what is commonly regarded as quality. It should be emphasized that in the process of selecting documents to be in- dexed by engines and in the ranking process as well, no human reviews are involved. But a certain bias can be found inherent in the ranking algo- rithms (Lewandowski, 2004b). These rate Web pages (apart from classic IR calculations) mainly by determining their popularity based on the link structure of the Web. The basic assumption is that a link to a page is a vote for that page. But not all links should be counted the same; link-based measures take into account the popularity of the linking page itself and the number of outgoing links, as well. This holds true for both of the main link-based ranking algorithms, Google’s PageRank (Page, Brin, Motwani, & Winograd, 1998) and HITS (Kleinberg, 1999). Table 1. Query-independent ranking factors (taken from Lewandowski, 2005a) Directory hierarchy Documents on a higher hierarchy level are preferred. Number of incoming links The higher the number of incoming links, the more important the document. Link popularity Quality/authority of a document is measured accord- ing to its linking within the Web graph. Click popularity Documents visited by many users are preferred. Up-to-dateness Current documents are preferred to older docu- ments. Document length Documents within a sudden length range are pre- ferred. File format Documents written in standard HTML are preferred to documents in other formats such as PDF or DOC. Size of the Website Documents from larger Web sites (or within a sud- den size range) are preferred. Link-based measures are commonly calculated query-independent, i.e., no computing power is needed to calculate these measures at the moment us- ers place their search queries. Therefore, these measures can be applied very fast by the ranking process. Other query-independent factors are used as well (see table 1 and for a detailed discussion Lewandowski, 2005a). 4 Here, the important point is that the process of ranking Web pages evolved from a query-document matching, based on term frequency and similar factors, to a process where several quality measurements are also taken into account. Link-based algorithms are of good use to push some highly relevant re- sults to the top of the results list. This approach is oriented towards the typical user behaviour. Users often view only a few results from the top of the list and seldom process to the second or even third page of the results list. Another prob- lem with the calculation of appropriate result lists is the shortness of search queries. Therefore, most ranking algorithms prefer popular pages and the presence of search terms in anchor texts. Although the general user rarely uses advanced search features, this does not make them unnecessary or useless. On the one hand, there are special user groups like librarians or in- formation professionals who conduct complex searches. On the other hand, while there is a majority of queries that can be successfully formulated without the use of advanced search syntax, one knows from his or her own searching behaviour that at least sometimes one needs to use operators or other advanced features.
Recommended publications
  • Risk Management Resolutions Easier Than a Diet; Good for the Health of Your Nonprofit by Melanie L
    A publication of the Nonprofit Risk Management Center Volume 14, No. 1, January/February 2005 Risk Management Resolutions Easier than a diet; good for the health of your nonprofit by Melanie L. Herman ere you one of thousands dilemma you’re facing and the of Americans who received a solution we recommend. To W gym membership gift tucked access this free service, visit neatly in a card from a loved one? www.nonprofitrisk.org and Across the country millions of click on the ADVICE tab. Or Americans are jotting down resolutions, give us a call at (202) 785-3891. most of which have something to do with the three Cs: calories, cardio- Resolution #1 workouts, or carcinogens. Blow the Dust off Resolutions about the three health- inducing Cs are awfully tough to keep, Your Policies as are risk management resolutions that Throughout the year the could change the health of your Nonprofit Risk Management nonprofit. Does the organization you Center receives phone calls from serve have a mission worth preserving? chief financial officers, Are there any risks lurking in your continued on page 2 nonprofit’s future that could spell significant set-back or disaster? Are you losing any sleep about risks related to HR, financial management, fundraising, NEW! Risk Management Essentials Series reputation, or staff/participant injuries? This article offers some simple but Visit www.nonprofitrisk.org to check out a important Risk Management special offer from the Nonprofit Risk Resolutions for 2005. The resolutions can be easily adapted to reflect the Management Center. The Risk Management circumstances and resources of your Essentials Series takes the guesswork out of nonprofit.
    [Show full text]
  • Uva-DARE (Digital Academic Repository)
    UvA-DARE (Digital Academic Repository) Search engine freedom: on the implications of the right to freedom of expression for the legal governance of Web search engines van Hoboken, J.V.J. Publication date 2012 Link to publication Citation for published version (APA): van Hoboken, J. V. J. (2012). Search engine freedom: on the implications of the right to freedom of expression for the legal governance of Web search engines. General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: https://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (https://dare.uva.nl) Download date:29 Sep 2021 Chapter 2: A Short History of Search Engines and Related Market Developments 22 2.1 The Internet, the Web and the rise of navigational media 2.1.1. Early visions of navigation in digitized information environments The way in which digital computing would lead to a revolution in information and knowledge navigation was already being explored more than half a century ago, when computers were still a rarity and neither the Internet, nor the World Wide Web existed.
    [Show full text]
  • Webcrawler: Finding What People Want
    © Copyright 2000 Brian Pinkerton WebCrawler: Finding What People Want Brian Pinkerton A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy University of Washington 2000 Program Authorized to Offer Degree: Department of Computer Science & Engineering University of Washington Graduate School This is to certify that I have examined this copy of a doctoral dissertation by Brian Pinkerton and have found that it is complete and satisfactory in all respects and that any and all revisions required by the final examining committee have been made. Co-chairs of the Supervisory Committee: _______________________________________________ Edward Lazowska _______________________________________________ John Zahorjan Reading Committee: _______________________________________________ Edward Lazowska _______________________________________________ John Zahorjan _______________________________________________ David Notkin Date: _____________________ In presenting this dissertation in partial fulfillment of the requirements for the Doctoral degree at the Univer- sity of Washington, I agree that the Library shall make its copies freely available for inspection. I further agree that extensive copying of the dissertation is allowable only for scholarly purposes, consistent with “fair use” as prescribed in the U.S. Copyright Law. Requests for copying or reproduction of this dissertation may be referred to Bell and Howell Information and Learning, 300 North Zeeb Road, Ann Arbor, MI 48106- 1346, to whom the author
    [Show full text]
  • The Commodification of Search
    San Jose State University SJSU ScholarWorks Master's Theses Master's Theses and Graduate Research 2008 The commodification of search Hsiao-Yin Chen San Jose State University Follow this and additional works at: https://scholarworks.sjsu.edu/etd_theses Recommended Citation Chen, Hsiao-Yin, "The commodification of search" (2008). Master's Theses. 3593. DOI: https://doi.org/10.31979/etd.wnaq-h6sz https://scholarworks.sjsu.edu/etd_theses/3593 This Thesis is brought to you for free and open access by the Master's Theses and Graduate Research at SJSU ScholarWorks. It has been accepted for inclusion in Master's Theses by an authorized administrator of SJSU ScholarWorks. For more information, please contact [email protected]. THE COMMODIFICATION OF SEARCH A Thesis Presented to The School of Journalism and Mass Communications San Jose State University In Partial Fulfillment of the Requirement for the Degree Master of Science by Hsiao-Yin Chen December 2008 UMI Number: 1463396 INFORMATION TO USERS The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleed-through, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note will indicate the deletion. ® UMI UMI Microform 1463396 Copyright 2009 by ProQuest LLC. All rights reserved. This microform edition is protected against unauthorized copying under Title 17, United States Code. ProQuest LLC 789 E.
    [Show full text]
  • Web Crawler for Mining Web Data
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395 -0056 Volume: 04 Issue: 02 | Feb -2017 www.irjet.net p-ISSN: 2395-0072 WEB CRAWLER FOR MINING WEB DATA S.AMUDHA, B.SC., M.SC., M.PHIL., Assistant Professor, VLB Janakiammal College of Arts and Science, Tamilnadu, India [email protected] -----------------------------------------------------------------********--------------------------------------------------------------------- ABSTRACT Web crawler is a computer program that browses the World Wide Web in a methodical, automated manner or in an orderly fashion. Web crawlers are full text search engines which assist users in navigating the web. Web crawling is an important method for collecting data on, and keeping up with, the rapidly expanding Internet. Users can find their resources by using different hypertext links. A vast number of web pages are continually being added every day, and information is constantly changing. Search engines are used to extract valuable Information from the internet. Web crawlers are the principal part of search engine, is a computer program or software that browses the World Wide Web in a methodical, automated manner or in an orderly fashion. This Paper is an overview of various types of Web Crawlers and the policies like selection, revisit, politeness, and parallelization. Key Words: Web Crawler, World Wide Web, Search Engine, Hyperlink, Uniform Resource Locator. 1. INTRODUCTION A Web crawler starts with a list of URLs to visit, called the seeds. As the crawler visits these URLs, it identifies all the hyperlinks in the page and adds them to the list of URLs to visit, called the crawl frontier. URLs from the frontier are recursively visited according to a set of policies.
    [Show full text]
  • Effective Performance of Information Retrieval on Web by Using Web Crawling
    International Journal of Web & Semantic Technology (IJWesT) Vol.3, No.2, April 2012 EFFECTIVE PERFORMANCE OF INFORMATION RETRIEVAL ON WEB BY USING WEB CRAWLING Sk.AbdulNabi 1 and Dr. P. Premchand 2 1Department of CSE, AVN Inst. of Engg. & Tech, Hyderabad, A.P., India [email protected] 2Department of CSE, Osmania University, Hyderabad, A.P., India [email protected] ABSTRACT World Wide Web consists of more than 50 billion pages online. It is highly dynamic [6] i.e. the web continuously introduces new capabilities and attracts many people. Due to this explosion in size, the effective information retrieval system or search engine can be used to access the information. In this paper we have proposed the EPOW (Effective Performance of WebCrawler) architecture. It is a software agent whose main objective is to minimize the overload of a user locating needed information. We have designed the web crawler by considering the parallelization policy. Since our EPOW crawler has a highly optimized system it can download a large number of pages per second while being robust against crashes. We have also proposed to use the data structure concepts for implementation of scheduler & circular Queue to improve the performance of our web crawler. (Abstract) KEYWORDS EPOW, Effective Web Crawler, Circular Queue, Scheduler, Basic Crawler, Precision & Recall. 1. INTRODUCTION In world wide web, information retrieval system [1] acts a vital role to extract the useful information with minimum amount of time. It has the ability to store the information and retrieved the information from the repository and also manages the information. The general objective of any Information Retrieval Method is to minimize the overhead of a user locating needed information.
    [Show full text]
  • Download Download
    International Journal of Management & Information Systems – Fourth Quarter 2011 Volume 15, Number 4 History Of Search Engines Tom Seymour, Minot State University, USA Dean Frantsvog, Minot State University, USA Satheesh Kumar, Minot State University, USA ABSTRACT As the number of sites on the Web increased in the mid-to-late 90s, search engines started appearing to help people find information quickly. Search engines developed business models to finance their services, such as pay per click programs offered by Open Text in 1996 and then Goto.com in 1998. Goto.com later changed its name to Overture in 2001, and was purchased by Yahoo! in 2003, and now offers paid search opportunities for advertisers through Yahoo! Search Marketing. Google also began to offer advertisements on search results pages in 2000 through the Google Ad Words program. By 2007, pay-per-click programs proved to be primary money-makers for search engines. In a market dominated by Google, in 2009 Yahoo! and Microsoft announced the intention to forge an alliance. The Yahoo! & Microsoft Search Alliance eventually received approval from regulators in the US and Europe in February 2010. Search engine optimization consultants expanded their offerings to help businesses learn about and use the advertising opportunities offered by search engines, and new agencies focusing primarily upon marketing and advertising through search engines emerged. The term "Search Engine Marketing" was proposed by Danny Sullivan in 2001 to cover the spectrum of activities involved in performing SEO, managing paid listings at the search engines, submitting sites to directories, and developing online marketing strategies for businesses, organizations, and individuals.
    [Show full text]
  • Indexing the World Wide Web: the Journey So Far Abhishek Das Google Inc., USA Ankit Jain Google Inc., USA
    Indexing The World Wide Web: The Journey So Far Abhishek Das Google Inc., USA Ankit Jain Google Inc., USA ABSTRACT In this chapter, we describe the key indexing components of today’s web search engines. As the World Wide Web has grown, the systems and methods for indexing have changed significantly. We present the data structures used, the features extracted, the infrastructure needed, and the options available for designing a brand new search engine. We highlight techniques that improve relevance of results, discuss trade-offs to best utilize machine resources, and cover distributed processing concept in this context. In particular, we delve into the topics of indexing phrases instead of terms, storage in memory vs. on disk, and data partitioning. We will finish with some thoughts on information organization for the newly emerging data-forms. INTRODUCTION The World Wide Web is considered to be the greatest breakthrough in telecommunications after the telephone. Quoting the new media reader from MIT press [Wardrip-Fruin , 2003]: “The World-Wide Web (W3) was developed to be a pool of human knowledge, and human culture, which would allow collaborators in remote sites to share their ideas and all aspects of a common project.” The last two decades have witnessed many significant attempts to make this knowledge “discoverable”. These attempts broadly fall into two categories: (1) classification of webpages in hierarchical categories (directory structure), championed by the likes of Yahoo! and Open Directory Project; (2) full-text index search engines such as Excite, AltaVista, and Google. The former is an intuitive method of arranging web pages, where subject-matter experts collect and annotate pages for each category, much like books are classified in a library.
    [Show full text]
  • Readingsample
    Information Science and Knowledge Management 14 Web Search Multidisciplinary Perspectives Bearbeitet von Amanda Spink, Michael Zimmer 1. Auflage 2008. Buch. xii, 352 S. Hardcover ISBN 978 3 540 75828 0 Format (B x L): 15,5 x 23,5 cm Gewicht: 703 g Weitere Fachgebiete > EDV, Informatik > EDV, Informatik: Allgemeines, Moderne Kommunikation > Soziale, sicherheitstechnische, ethische Aspekte Zu Inhaltsverzeichnis schnell und portofrei erhältlich bei Die Online-Fachbuchhandlung beck-shop.de ist spezialisiert auf Fachbücher, insbesondere Recht, Steuern und Wirtschaft. Im Sortiment finden Sie alle Medien (Bücher, Zeitschriften, CDs, eBooks, etc.) aller Verlage. Ergänzt wird das Programm durch Services wie Neuerscheinungsdienst oder Zusammenstellungen von Büchern zu Sonderpreisen. Der Shop führt mehr als 8 Millionen Produkte. 16 Web Searching: A Quality Measurement Perspective D. Lewandowski and N. Höchstötter Summary The purpose of this paper is to describe various quality measures for search engines and to ask whether these are suitable. We especially focus on user needs and their use of Web search engines. The paper presents an extensive litera- ture review and a first quality measurement model, as well. Findings include that Web search engine quality can not be measured by just retrieval effectiveness (the quality of the results), but should also consider index quality, the quality of the search features and Web search engine usability . For each of these sections, empiri- cal results from studies conducted in the past, as well as from our own research are presented. These results have implications for the evaluation of Web search engines and for the development of better search systems that give the user the best possible search experience.
    [Show full text]
  • Intelligent Web Agent for Search Engines
    International Conference on Trends and Advances in Computation and Engineering, TRACE- 2010 Intelligent Web Agent for Search Engines Avinash N. Bhute1,Harsha A. Bhute2 , Dr.B.B.Meshram3 Abstract II. WEB CRAWLER In this paper we review studies of the growth of the Internet and technologies that are useful for information search and retrieval on A Web crawler is a computer program that browses the the Web. Search engines are retrieve the efficient information. We World Wide Web in a methodical, automated manner. Other collected data on the Internet from several different sources, e.g., terms for Web crawlers are ants, automatic indexers, bots, and current as well as projected number of users, hosts, and Web sites. worm [3] or Web spider, Web robot. Web crawling or The trends cited by the sources are consistent and point to spidering is term alternatively used for same. Many sites, in exponential growth in the past and in the coming decade. Hence it is particular search engines, use spidering as a means of not surprising that about 85% of Internet users surveyed claim using search engines and search services to find specific information and providing up-to-date data. Web crawlers are mainly used to users are not satisfied with the performance of the current generation create a copy of all the visited pages for later processing by a of search engines; the slow retrieval speed, communication delays, search engine that will index the downloaded pages to provide and poor quality of retrieved results. Web agents, programs acting fast searches. Crawlers can also be used for automating autonomously on some task, are already present in the form of maintenance tasks on a Web site, such as checking links or spiders, crawler, and robots.
    [Show full text]
  • Looksmart Expands Strategic Relationship with Infospace, Bringing Its Search Results to Excite and Webcrawler
    LookSmart Expands Strategic Relationship With InfoSpace, Bringing Its Search Results to Excite and Webcrawler SAN FRANCISCO & BELLEVUE, Wash.--(BUSINESS WIRE)--May 7, 2002-- LookSmart's Search Results, Found Within InfoSpace's Next Generation Meta-Search Product, Launches Today at Excite.com and WebCrawler.com LookSmart (Nasdaq:LOOK; ASX:LOK), a global leader in search targeted marketing, and InfoSpace, Inc. (Nasdaq:INSP), a provider of wireless and Internet software and application services, today announced that they have extended and expanded their search relationship. Under the agreement, LookSmart search listings will now appear on InfoSpace's newly-launched, next generation, meta-search product found at Excite (www.excite.com) and WebCrawler (www.webcrawler.com). Excite is one of the leading personalization portals on the Internet, offering world-class content and functionality from more than 75 leading sources with more than 14 million monthly unique users. Additionally, LookSmart will continue to provide results for InfoSpace's other meta-search solutions, which include Dogpile (www.dogpile.com) and MetaCrawler (www.metacrawler.com). InfoSpace's proprietary meta-search technology allows users to search multiple engines at once, returning comprehensive results fast. "LookSmart is pleased to expand its relationship with InfoSpace by playing a significant role in the new search product that is served on Excite and WebCrawler," said Scott Stanford, senior vice president of business development for LookSmart. "LookSmart will meet an immediate need for high-quality search results and we look forward to continue working together to create great search for users of InfoSpace's properties." "LookSmart's search listings produce highly relevant results for users of our meta-search products, including the next generation engine launched at Excite and WebCrawler," said York Baur, InfoSpace executive vice president, wireline and broadband.
    [Show full text]
  • Effective Searching Policies for Web Crawler
    International Journal of Modern Engineering Research (IJMER) www.ijmer.com Vol. 3, Issue. 5, Sep - Oct. 2013 pp-3137-3139 ISSN: 2249-6645 Effective Searching Policies for Web Crawler Suyash Gupta, KrishanDev Mishra, Prerna Singh (Department of Information Technology, Mahamaya Technical University, India) (Department of Information Technology, Mahamaya Technical University, India) (Department of Information Technology, Mahamaya Technical University, India) ABSTRACT: As we know search engines cannot index every Web page, due to limited storage, bandwidth, computational resources and the dynamic nature of the web. It cannot monitored continuously all parts of the web for changes. Therefore it is important to develop effective searching policies. In this technique there is the combination of different searching technique to form a effective searching policies. These combined techniques are best first search, focused crawling, info spiders, recrawling pages for updation, crawling the hidden web page.This combined technique also includes Selection policy such as page rank, path ascending, focused crawling Revisit policy such as freshness , age Politeness , Parallelization so that it allow distributed web crawling. Keywords: Best first search, recrawling pages for updation, crawling the hidden web page, focused crawling, Revisit policy, Politeness, Parallelization. I. INTRODUCTION A Web crawler is a software program that helps in locating information stored on a computer system, typically based on WWW. Web crawlers are mainly used to create a copy of all the visited pages for later processing by a search engine that will index the downloaded pages to provide fast searches. Crawlers can also be used for automating maintenance tasks on a Web site, such as checking links or validating HTML code.
    [Show full text]