Heritrix Documentation

Total Page:16

File Type:pdf, Size:1020Kb

Heritrix Documentation Heritrix Documentation Internet Archive and contributors Sep 23, 2021 Contents: 1 Getting Started with Heritrix3 1.1 System Requirements..........................................3 1.2 Installation................................................3 1.3 Environment Variables..........................................3 1.4 Runnning Heritrix............................................4 1.5 Accessing the User Interface.......................................4 1.6 Your First Crawl.............................................4 1.7 Exiting Heritrix..............................................5 2 Operating Heritrix 7 2.1 Running Heritrix.............................................7 2.1.1 Command-line Options.....................................7 2.1.2 Environment Variables.....................................8 2.2 Security Considerations.........................................8 2.2.1 Understanding the Risks....................................8 2.2.2 Network Access Control....................................8 2.2.3 Login Authentication Access Control.............................9 2.3 Log Files.................................................9 2.3.1 alerts.log............................................9 2.3.2 crawl.log............................................9 2.3.3 progress-statistics.log...................................... 10 2.3.4 runtime-errors.log........................................ 11 2.3.5 uri-errors.log.......................................... 11 2.4 Reports.................................................. 11 2.4.1 Crawl Summary (crawl-report.txt)............................... 11 2.4.2 Seeds (seeds-report.txt)..................................... 12 2.4.3 Hosts (hosts-report.txt)..................................... 12 2.4.4 SourceTags (source-report.txt)................................. 13 2.4.5 Mimetypes (mimetype-report.txt)............................... 13 2.4.6 ResponseCode (responsecode-report.txt)............................ 14 2.4.7 Processors (processors-report.txt)............................... 14 2.4.8 FrontierSummary (frontier-summary-report.txt)........................ 15 2.4.9 ToeThreads (threads-report.txt)................................. 15 2.5 Action Directory............................................. 15 2.6 Crawl Recovery............................................. 16 2.6.1 Full recovery.......................................... 16 i 2.6.2 Split Recovery......................................... 17 3 Configuring Crawl Jobs 19 3.1 Basic Job Settings............................................ 19 3.1.1 Crawl Limits.......................................... 19 3.1.2 maxToeThreads......................................... 20 3.1.3 metadata.operatorContactUrl.................................. 20 3.1.4 Robots.txt Honoring Policy................................... 20 3.2 Crawl Scope............................................... 20 3.2.1 Decide Rules.......................................... 21 3.2.2 DecideRuleSequence Logging................................. 23 3.3 Frontier.................................................. 23 3.3.1 Politeness............................................ 23 3.3.2 Retry Policy........................................... 23 3.3.3 Bandwidth Limits........................................ 24 3.3.4 Extractor Parameters...................................... 24 3.4 Sheets (Site-specific Settings)...................................... 24 3.5 Other Protocols.............................................. 25 3.5.1 FTP............................................... 26 3.5.2 SFTP.............................................. 26 3.5.3 WHOIS............................................. 27 3.6 Modifying a Running Job........................................ 27 3.6.1 Browse Beans.......................................... 28 3.6.2 Scripting Console........................................ 29 4 Bean Reference 31 4.1 Core Beans................................................ 31 4.1.1 ActionDirectory......................................... 31 4.1.2 BdbCookieStore........................................ 32 4.1.3 BdbFrontier........................................... 32 4.1.4 BdbModule........................................... 32 4.1.5 BdbServerCache........................................ 33 4.1.6 BdbUriUniqFilter........................................ 33 4.1.7 CheckpointService....................................... 34 4.1.8 CrawlController......................................... 34 4.1.9 CrawlerLoggerModule..................................... 35 4.1.10 CrawlLimitEnforcer...................................... 36 4.1.11 CrawlMetadata......................................... 37 4.1.12 CredentialStore......................................... 37 4.1.13 DiskSpaceMonitor....................................... 38 4.1.14 RulesCanonicalizationPolicy.................................. 38 4.1.15 SheetOverlaysManager..................................... 38 4.1.16 StatisticsTracker........................................ 39 4.1.17 TextSeedModule........................................ 40 4.2 Decide Rules............................................... 40 4.2.1 AcceptDecideRule....................................... 40 4.2.2 ClassKeyMatchesRegexDecideRule.............................. 40 4.2.3 ContentLengthDecideRule................................... 41 4.2.4 ContentTypeMatchesRegexDecideRule............................ 41 4.2.5 ContentTypeNotMatchesRegexDecideRule.......................... 41 4.2.6 ExpressionDecideRule (contrib)................................ 41 4.2.7 ExternalGeoLocationDecideRule................................ 41 4.2.8 FetchStatusDecideRule..................................... 42 4.2.9 FetchStatusMatchesRegexDecideRule............................. 42 ii 4.2.10 FetchStatusNotMatchesRegexDecideRule........................... 42 4.2.11 HasViaDecideRule....................................... 42 4.2.12 HopCrossesAssignmentLevelDomainDecideRule....................... 42 4.2.13 HopsPathMatchesRegexDecideRule.............................. 43 4.2.14 IdenticalDigestDecideRule................................... 43 4.2.15 IpAddressSetDecideRule.................................... 43 4.2.16 MatchesFilePatternDecideRule................................. 44 4.2.17 MatchesListRegexDecideRule................................. 44 4.2.18 MatchesRegexDecideRule................................... 44 4.2.19 MatchesStatusCodeDecideRule................................ 44 4.2.20 NotMatchesFilePatternDecideRule............................... 45 4.2.21 NotMatchesListRegexDecideRule............................... 45 4.2.22 NotMatchesRegexDecideRule................................. 45 4.2.23 NotMatchesStatusCodeDecideRule.............................. 45 4.2.24 NotOnDomainsDecideRule................................... 45 4.2.25 NotOnHostsDecideRule.................................... 46 4.2.26 NotSurtPrefixedDecideRule.................................. 46 4.2.27 OnDomainsDecideRule..................................... 46 4.2.28 OnHostsDecideRule...................................... 46 4.2.29 PathologicalPathDecideRule.................................. 46 4.2.30 PredicatedDecideRule..................................... 47 4.2.31 PrerequisiteAcceptDecideRule................................. 47 4.2.32 RejectDecideRule........................................ 47 4.2.33 ResourceLongerThanDecideRule................................ 47 4.2.34 ResourceNoLongerThanDecideRule.............................. 48 4.2.35 ResponseContentLengthDecideRule.............................. 48 4.2.36 SchemeNotInSetDecideRule.................................. 48 4.2.37 ScriptedDecideRule....................................... 48 4.2.38 SeedAcceptDecideRule..................................... 49 4.2.39 SourceSeedDecideRule..................................... 49 4.2.40 SurtPrefixedDecideRule.................................... 49 4.2.41 TooManyHopsDecideRule................................... 50 4.2.42 TooManyPathSegmentsDecideRule.............................. 50 4.2.43 TransclusionDecideRule.................................... 50 4.2.44 ViaSurtPrefixedDecideRule................................... 51 4.3 Candidate Processors........................................... 51 4.3.1 CandidateScoper........................................ 51 4.3.2 FrontierPreparer......................................... 51 4.4 Pre-Fetch Processors........................................... 52 4.4.1 PreconditionEnforcer...................................... 52 4.4.2 Preselector........................................... 52 4.5 Fetch Processors............................................. 53 4.5.1 FetchDNS............................................ 53 4.5.2 FetchFTP............................................ 53 4.5.3 FetchHTTP........................................... 54 4.5.4 FetchSFTP........................................... 56 4.5.5 FetchWhois........................................... 57 4.6 Link Extractors.............................................. 58 4.6.1 ExtractorChrome (contrib)................................... 58 4.6.2 ExtractorCSS.......................................... 59 4.6.3 ExtractorDOC.......................................... 59 4.6.4 ExtractorHTML......................................... 59 4.6.5 AggressiveExtractorHTML................................... 60 4.6.6 JerichoExtractorHTML..................................... 60 iii 4.6.7 ExtractorHTMLForms..................................... 61 4.6.8 ExtractorHTTP........................................
Recommended publications
  • Harvesting Strategies for a National Domain France Lasfargues, Clément Oury, Bert Wendland
    Legal deposit of the French Web: harvesting strategies for a national domain France Lasfargues, Clément Oury, Bert Wendland To cite this version: France Lasfargues, Clément Oury, Bert Wendland. Legal deposit of the French Web: harvesting strategies for a national domain. International Web Archiving Workshop, Sep 2008, Aarhus, Denmark. hal-01098538 HAL Id: hal-01098538 https://hal-bnf.archives-ouvertes.fr/hal-01098538 Submitted on 26 Dec 2014 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Distributed under a Creative Commons Attribution| 4.0 International License Legal deposit of the French Web: harvesting strategies for a national domain France Lasfargues, Clément Oury, and Bert Wendland Bibliothèque nationale de France Quai François Mauriac 75706 Paris Cedex 13 {france.lasfargues, clement.oury, bert.wendland}@bnf.fr ABSTRACT 1. THE FRENCH CONTEXT According to French Copyright Law voted on August 1st, 2006, the Bibliothèque nationale de France (“BnF”, or “the Library”) is 1.1 Defining the scope of the legal deposit in charge of collecting and preserving the French Internet. The On August 1st, 2006, a new Copyright law was voted by the Library has established a “mixed model” of Web archiving, which French Parliament.
    [Show full text]
  • Web Archiving Environmental Scan
    Web Archiving Environmental Scan The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Truman, Gail. 2016. Web Archiving Environmental Scan. Harvard Library Report. Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:25658314 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA Web Archiving Environmental Scan Harvard Library Report January 2016 Prepared by Gail Truman The Harvard Library Report “Web Archiving Environmental Scan” is licensed under a Creative Commons Attribution 4.0 International License. Prepared by Gail Truman, Truman Technologies Reviewed by Andrea Goethals, Harvard Library and Abigail Bordeaux, Library Technology Services, Harvard University Revised by Andrea Goethals in July 2017 to correct the number of dedicated web archiving staff at the Danish Royal Library This report was produced with the generous support of the Arcadia Fund. Citation: Truman, Gail. 2016. Web Archiving Environmental Scan. Harvard Library Report. Table of Contents Executive Summary ............................................................................................................................ 3 Introduction ......................................................................................................................................
    [Show full text]
  • User Manual [Pdf]
    Heritrix User Manual Internet Archive Kristinn Sigur#sson Michael Stack Igor Ranitovic Table of Contents 1. Introduction ............................................................................................................ 1 2. Installing and running Heritrix .................................................................................... 2 2.1. Obtaining and installing Heritrix ...................................................................... 2 2.2. Running Heritrix ........................................................................................... 3 2.3. Security Considerations .................................................................................. 7 3. Web based user interface ........................................................................................... 7 4. A quick guide to running your first crawl job ................................................................ 8 5. Creating jobs and profiles .......................................................................................... 9 5.1. Crawl job .....................................................................................................9 5.2. Profile ....................................................................................................... 10 6. Configuring jobs and profiles ................................................................................... 11 6.1. Modules (Scope, Frontier, and Processors) ....................................................... 12 6.2. Submodules ..............................................................................................
    [Show full text]
  • Web Archiving for Academic Institutions
    University of San Diego Digital USD Digital Initiatives Symposium Apr 23rd, 1:00 PM - 4:00 PM Web Archiving for Academic Institutions Lori Donovan Internet Archive Mary Haberle Internet Archive Follow this and additional works at: https://digital.sandiego.edu/symposium Donovan, Lori and Haberle, Mary, "Web Archiving for Academic Institutions" (2018). Digital Initiatives Symposium. 4. https://digital.sandiego.edu/symposium/2018/2018/4 This Workshop is brought to you for free and open access by Digital USD. It has been accepted for inclusion in Digital Initiatives Symposium by an authorized administrator of Digital USD. For more information, please contact [email protected]. Web Archiving for Academic Institutions Presenter 1 Title Senior Program Manager, Archive-It Presenter 2 Title Web Archivist Session Type Workshop Abstract With the advent of the internet, content that institutional archivists once preserved in physical formats is now web-based, and new avenues for information sharing, interaction and record-keeping are fundamentally changing how the history of the 21st century will be studied. Due to the transient nature of web content, much of this information is at risk. This half-day workshop will cover the basics of web archiving, help attendees identify content of interest to them and their communities, and give them an opportunity to interact with tools that assist with the capture and preservation of web content. Attendees will gain hands-on web archiving skills, insights into selection and collecting policies for web archives and how to apply what they've learned in the workshop to their own organizations. Location KIPJ Room B Comments Lori Donovan works with partners and the Internet Archive’s web archivists and engineering team to develop the Archive-It service so that it meets the needs of memory institutions.
    [Show full text]
  • Getting Started in Web Archiving
    Submitted on: 13.06.2017 Getting Started in Web Archiving Abigail Grotke Library Services, Digital Collections Management and Services Division, Library of Congress, Washington, D.C., United States. E-mail address: [email protected] This work is made available under the terms of the Creative Commons Attribution 4.0 International License: http://creativecommons.org/licenses/by/4.0 Abstract: This purpose of this paper is to provide general information about how organizations can get started in web archiving, for both those who are developing new web archiving programs and for libraries that are just beginning to explore the possibilities. The paper includes an overview of considerations when establishing a web archiving program, including typical approaches that national libraries take when preserving the web. These include: collection development, legal issues, tools and approaches, staffing, and whether to do work in-house or outsource some or most of the work. The paper will introduce the International Internet Preservation Consortium and the benefits of collaboration when building web archives. Keywords: web archiving, legal deposit, collaboration 1 BACKGROUND In the more than twenty five years since the World Wide Web was invented, it has been woven into everyday life—a platform by which a huge number of individuals and more traditional publishers distribute information and communicate with one another around the world. While the size of the web can be difficult to articulate, it is generally understood that it is large and ever-changing and that content is continuously added and removed. With so much global cultural heritage being documented online, librarians, archivists, and others are increasingly becoming aware of the need to preserve this valuable resource for future generations.
    [Show full text]
  • Incremental Crawling with Heritrix
    Incremental crawling with Heritrix Kristinn Sigurðsson National and University Library of Iceland Arngrímsgötu 3 107 Reykjavík Iceland [email protected] Abstract. The Heritrix web crawler aims to be the world's first open source, extensible, web-scale, archival-quality web crawler. It has however been limited in its crawling strategies to snapshot crawling. This paper reports on work to add the ability to conduct incremental crawls to its capabilities. We first discuss the concept of incremental crawling as opposed to snapshot crawling and then the possible ways to design an effective incremental strategy. An overview is given of the implementation that we did, its limits and strengths are discussed. We then report on the results of initial experimentation with the new software which have gone well. Finally, we discuss issues that remain unresolved and possible future improvements. 1 Introduction With an increasing number of parties interested in crawling the World Wide Web, for a variety of reasons, a number of different crawl types have emerged. The development team at Internet Archive [12] responsible for the Heritrix web crawler, have highlighted four distinct variations [1], broad , focused, continuous and experimental crawling. Broad and focused crawls are in many ways similar, the primary difference being that broad crawls emphasize capturing a large scope1, whereas focused crawling calls for a more complete coverage of a smaller scope. Both approaches use a snapshot strategy , which involves crawling the scope once and once only. Of course, crawls are repeatable but only by starting again from the seeds. No information from past crawls is used in new ones, except possibly some changes to the configuration made by the operator, to avoid crawler traps etc.
    [Show full text]
  • Adaptive Revisiting with Heritrix Master Thesis (30 Credits/60 ECTS)
    University of Iceland Faculty of Engineering Department of Computer Science Adaptive Revisiting with Heritrix Master Thesis (30 credits/60 ECTS) by Kristinn Sigurðsson May 2005 Supervisors: Helgi Þorbergsson, PhD Þorsteinn Hallgrímsson Útdráttur á íslensku Veraldarvefurinn geymir sívaxandi hluta af þekkingu og menningararfi heimsins. Þar sem Vefurinn er einnig sífellt að breytast þá er nú unnið ötullega að því að varðveita innihald hans á hverjum tíma. Þessi vinna er framlenging á skylduskila lögum sem hafa í síðustu aldir stuðlað að því að varðveita prentað efni. Fyrstu þrír kaflarnir lýsa grundvallar erfiðleikum við það að safna Vefnum og kynnir hugbúnaðinn Heritrix, sem var smíðaður til að vinna það verk. Fyrsti kaflinn einbeitir sér að ástæðunum og bakgrunni þessarar vinnu en kaflar tvö og þrjú beina kastljósinu að tæknilegri þáttum. Markmið verkefnisins var að þróa nýja tækni til að safna ákveðnum hluta af Vefnum sem er álitinn breytast ört og vera í eðli sínu áhugaverður. Seinni kaflar fjalla um skilgreininu á slíkri aðferðafræði og hvernig hún var útfærð í Heritrix. Hluti þessarar umfjöllunar beinist að því hvernig greina má breytingar í skjölum. Að lokum er fjallað um fyrstu reynslu af nýja hugbúnaðinum og sjónum er beint að þeim þáttum sem þarfnast frekari vinnu eða athygli. Þar sem markmiðið með verkefninu var að leggja grunnlínur fyrir svona aðferðafræði og útbúa einfalda og stöðuga útfærsla þá inniheldur þessi hluti margar hugmyndir um hvað mætti gera betur. Keywords Web crawling, web archiving, Heritrix, Internet, World Wide Web, legal deposit, electronic legal deposit. i Abstract The World Wide Web contains an increasingly significant amount of the world’s knowledge and heritage.
    [Show full text]
  • An Introduction to Heritrix an Open Source Archival Quality Web Crawler
    An Introduction to Heritrix An open source archival quality web crawler Gordon Mohr, Michael Stack, Igor Ranitovic, Dan Avery and Michele Kimpton Internet Archive Web Team {gordon,stack,igor,dan,michele}@archive.org Abstract. Heritrix is the Internet Archive's open-source, extensible, web-scale, archival-quality webcrawler project. The Internet Archive started Heritrix development in the early part of 2003. The intention was to develop a crawler for the specific purpose of archiving websites and to support multiple different use cases including focused and broadcrawling. The software is open source to encourage collaboration and joint development across institutions with similar needs. A pluggable, extensible architecture facilitates customization and outside contribution. Now, after over a year of development, the Internet Archive and other institutions are using Heritrix to perform focused and increasingly broad crawls. Introduction The Internet Archive (IA) is a 5013C non-profit corporation, whose mission is to build a public Internet digital library. Over the last 6 years, IA has built the largest public web archive to date, hosting over 400 TB of data. The Web Archive is comprised primarily of pages collected by Alexa Internet starting in 1996. Alexa Internet is a Web cataloguing company founded by Brewster Kahle and Bruce Gilliat in 1996. Alexa Internet takes a snapshot of the web every 2 months, currently collecting 10 TB of data per month from over 35 million sites. Alexa Internet donates this crawl to the Internet Archive, and IA stores and indexes the collection. Alexa uses its own proprietary software and techniques to crawl the web. This software is not available to Internet Archive or other institutions for use or extension.
    [Show full text]
  • Exploring the Intersections of Web Science and Accessibility 3
    Exploring the Intersections of Web Science and Accessibility TREVOR BOSTIC, JEFF STANLEY, JOHN HIGGINS, DANIEL CHUDNOV, RACHAEL L. BRADLEY MONTGOMERY, JUSTIN F. BRUNELLE, The MITRE Corporation The web is the prominent way information is exchanged in the 21st century. However, ensuring web-based information is accessible is complicated, particularly with web applications that rely on JavaScript and other technologies to deliver and build representations; representations are often the HTML, images, or other code a server delivers for a web resource. Static representations are becoming rarer and assessing the accessibility of web-based information to ensure it is available to all users is increasingly difficult given the dynamic nature of representations. In this work, we survey three ongoing research threads that can inform web accessibility solutions: assessing web accessibility, modeling web user activity, and web application crawling. Current web accessibility research is contin- ually focused on increasing the percentage of automatically testable standards, but still relies heavily upon manual testing for complex interactive applications. Along-side web accessibility research, there are mechanisms developed by researchers that replicate user interactions with web pages based on usage patterns. Crawling web applications is a broad research domain; exposing content in web applications is difficult because of incompatibilities in web crawlers and the technologies used to create the applications. We describe research on crawling the deep web by exercising user forms. We close with a thought exercise regarding the convergence of these three threads and the future of automated, web-based accessibility evaluation and assurance through a use case in web archiving. These research efforts provide insight into how users interact with websites, how to automate and simulate user interactions, how to record the results of user interactions, and how to analyze, evaluate, and map resulting website content to determine its relative accessibility.
    [Show full text]
  • Partnership Opportunities with the Internet Archive Web Archiving in Libraries October 21, 2020
    Partnership Opportunities with the Internet Archive Web archiving in libraries October 21, 2020 Karl-Rainer Blumenthal Web Archivist, Internet Archive Web archiving is the process of collecting, preserving, and enabling access to web-published materials. Average lifespan of a webpage 92 days WEB ARCHIVING crawler replay app W/ARC WEB ARCHIVING TECHNOLOGY Brozzler Heritrix ARC HTTrack WARC warcprox wget Wayback Machine OpenWayback pywb wab.ac oldweb.today WEB ARCHIVING TECHNOLOGY Brozzler Heritrix ARC HTTrack WARC warcprox wget Archive-It Wayback Machine NetarchiveSuite (DK/FR) OpenWayback PANDAS (AUS) pywb Web Curator (UK/NZ) wab.ac Webrecorder oldweb.today WEB ARCHIVING The Wayback Machine The largest publicly available web archive in existence. https://archive.org/web/ > 300 Billion Web Pages > 100 million websites > 150 languages ~ 1 billion URLs added per week WEB ARCHIVING The Wayback Machine The largest publicly available web archive in existence. https://archive.org/web/ > 300 Billion Web Pages > 100 million websites > 150 languages ~ 1 billion URLs added per week WEB ARCHIVING The Wayback Machine Limitations: Lightly curated Completeness Temporal cohesion Access: No full-text search No descriptive metadata Access by URL only ARCHIVE-IT Archive-It https://archive-it.org Curator controlled > 700 partner organizations ~ 2 PB of web data collected Full text and metadata searchable APIs for archives, metadata, search, &c. ARCHIVE-IT COLLECTIONS ARCHIVE-IT PARTNERS WEB ARCHIVES AS DATA WEB ARCHIVES AS DATA WEB ARCHIVES AS DATA WEB ARCHIVES AS DATA WEB ARCHIVES AS (GOV) DATA WEB ARCHIVE ACCESSIBILITY WEB ARCHIVE ACCESSIBILITY WEB ARCHIVING COLLABORATION FDLP Libraries WEB ARCHIVING COLLABORATION FDLP Libraries Archive-It partners THANKS <3 ...and keep in touch! Karl-Rainer Blumenthal Web Archivist, Internet Archive [email protected] [email protected] Partnership Opportunities with Internet Archive Andrea Mills – Digitization Program Manager, Internet Archive 1.
    [Show full text]
  • Collecte Orientée Sur Le Web Pour La Recherche D'information Spécialisée
    Collecte orientée sur le Web pour la recherche d’information spécialisée Clément de Groc To cite this version: Clément de Groc. Collecte orientée sur le Web pour la recherche d’information spécialisée. Autre [cs.OH]. Université Paris Sud - Paris XI, 2013. Français. NNT : 2013PA112073. tel-00853250 HAL Id: tel-00853250 https://tel.archives-ouvertes.fr/tel-00853250 Submitted on 22 Aug 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Université Paris-Sud École doctorale d’informatique LIMSI-CNRS Collecte orientée sur le Web pour la recherche d’information spécialisée par Clément de Groc Thèse présentée pour obtenir le grade de Docteur de l’Université Paris-Sud Soutenue publiquement à Orsay le mercredi 5 juin 2013 en présence d’un jury composé de : Rapporteur Éric Gaussier Université de Grenoble Rapporteur Jacques Savoy Université de Neuchâtel Examinatrice Chantal Reynaud Université Paris-Sud Examinateur Mohand Boughanem Université de Toulouse Directeur Pierre Zweigenbaum CNRS Co-directeur Xavier Tannier Université Paris-Sud Invité Claude de Loupy Syllabs REMERCIEMENTS Cette thèse est l’aboutissement d’un travail individuel qui a nécessité le soutien et la participation de nombreuses personnes. Je remercie messieurs Éric Gaussier et Jacques Savoy de m’avoir fait l’immense honneur d’accep- ter d’être rapporteurs de mon travail, au même titre que madame Chantal Reynaud et monsieur Mohand Boughanem qui ont accepté de faire partie de mon jury.
    [Show full text]
  • The Use Case of Heritrix?
    An Architecture for Selective Web Harvesting: The Use Case of Heritrix? Vassilis Plachouras1, Florent Carpentier2, Julien Masan`es2, Thomas Risse3, Pierre Senellart4, Patrick Siehndel3, and Yannis Stavrakas1 1 IMIS, ATHENA Research Center, Athens, Greece 2 Internet Memory Foundation, 93100 Montreuil, France 3 L3S Research Center, University of Hannover, Germany 4 Institut Mines-T´el´ecom; T´el´ecomParisTech; CNRS LTCI, Paris, France Abstract. In this paper we provide a brief overview of the crawling architecture of ARCOMEM and how it addresses the challenges arising in the context of selective web harvesting. We describe some of the main technologies developed to perform selective harvesting and we focus on a modified version of the open source crawler Heritrix, which we have adapted to fit in ACROMEM's crawling architecture. The simulation experiments we have performed show that the proposed architecture is effective in a focused crawling setting. Keywords: Web harvesting, focused crawling, web application crawling 1 Introduction The objectives of the ARCOMEM project are to select and harvest the web content that best preserves community memories for the future by exploiting social media, as opposed to exhaustive web harvesting approaches. To achieve these objectives, there are several challenges that need to be addressed. First, it is necessary to integrate in the crawling process advanced and computationally- intensive content analysis without reducing the efficiency of the crawler. Second, it is important to guide the web harvesting with accuracy to discover and collect relevant resources. Last, the above functionalities must be scalable in order to process large amounts of data. In this paper, we provide a brief description of the ARCOMEM's crawling architecture and some of the main technologies.
    [Show full text]