Making Recommendations from Web Archives for “Lost” Web Pages

Making Recommendations from Web Archives for “Lost” Web Pages

Making Recommendations from Web Archives for “Lost” Web Pages Lulwah M. Alkwai, Michael L. Nelson, Michele C. Weigle Department of Computer Science Old Dominion University Norfolk, Virginia 23529 USA [email protected],{mln,mweigle}@cs.odu.edu ABSTRACT enhance the response from a web archive with recommendations of When a user requests a web page from a web archive, the user will other archived web pages that may be relevant to the request. For typically either get an HTTP 200 if the page is available, or an HTTP example, Figure 1b shows a potential set of recommended archived 404 if the web page has not been archived. This is because web web pages for the request in Figure 1a. archives are typically accessed by Uniform Resource Identifier (URI) One approach to finding related web pages is to examine the lookup, and the response is binary: the archive either has the page content of the requested web page and then select candidates with or it does not, and the user will not know of other archived web similar content. However, in this work, we assume that the re- pages that exist and are potentially similar to the requested web quested web page is neither available in web archives nor on the page. In this paper, we propose augmenting these binary responses live web and thus is considered to be a “lost” web page. This as- with a model for selecting and ranking recommended web pages sumption reflects previous work showing that users often search in a Web archive. This is to enhance both HTTP 404 responses web archives when they cannot find the desired web page on the and HTTP 200 responses by surfacing web pages in the archive live web [5] and that there are a significant number of web pages that the user may not know existed. First, we check if the URI is that are not archived [1, 3]. Learning about a requested web page already classified in DMOZ or Wikipedia. If the requested URI isnot without examining the content of the page can be challenging due found, we use machine learning to classify the URI using DMOZ to little context and content available. There are several advan- as our ontology and collect candidate URIs to recommended to the tages to using the Uniform Resource Identifier (URI) over using user. The classification is in two parts, a first-level classification the content of the web page. First, in some cases the content of the and a deep classification. Next, we filter the candidates based onif URI is not available on the live Web or in the archive. Second, the they are present in the archive. Finally, we rank candidates based URI may contain hints about the resource it identifies. Third, it is on several features, such as archival quality, web page popularity, more efficient both in time and space to use the text oftheURI only rather than to extract the content of the web page. Fourth, temporal similarity, and URI similarity. We calculated the F1 score for different methods of classifying the requested web page atthe some web pages have little or no textual content, such as images or first level. We found that using all-grams from the URI after remov- videos, so extracting the content will be not useful or even possible. ing numerals and the top-level domain (TLD) produced the best Fifth, some web pages have privacy settings that do not permit them to be archived. result with F1=0.59. For the deep-level classification, we measured the accuracy at each classification level. For second-level classifica- In this work we recommend similar URIs to a request by follow- ing five steps. First, we determine if the requested URI isoneof tion, the micro-average F1=0.30 and for third-level classification, the 4 million categorized URIs in DMOZ1 or in Wikipedia via the F1=0.15. We also found that 44.89% of the correctly classified URIs contained at least one word that exists in a dictionary and 50.07% of Wikipedia API. If the URI is found, we collect candidates in the same the correctly classified URIs contained long strings in the domain. category from DMOZ or Wikipedia and move to Step 4. Second, In comparison with the URIs from our Wayback access logs, only if the URI is not found we classify the requested URI based on a first-level of categorization. Third, we classify the requested URI arXiv:1908.02819v1 [cs.DL] 7 Aug 2019 5.39% of those URIs contained only words from a dictionary, and 26.74% contained at least one word from a dictionary. These per- to determine the deep categorization levels and collect candidates. centages are low and may affect the ability for the requested URI Fourth, we filter candidates by removing candidates that are not to be correctly classified. archived. Finally, we filter and rank candidates based on several features, such as archival quality, web page popularity, temporal similarity, and URI similarity. 1 INTRODUCTION Web archives are a window to view past versions of web pages. 2 RELATED WORK The oldest and largest web archive, the Internet Archive’s Wayback There has been previous work on searching an archive without in- Machine, contains over 700 billion web objects [16] . But even with dexing it. Kanhabua et al. [19] proposed a search system to support this massive collection, sometimes a user requests a web page that retrieval and analytics on the Internet Archive. They used Bing to the Wayback Machine does not have. Currently, in this case, the search the live web and then extracted the URLs from the results user is presented with a message saying that the Wayback Machine does not have the page archived and a link to search for other 1The original DMOZ, http://dmoz.org, is out of service but we have archived versions archived pages in that same domain (Figure 1a). Our goal is to locally. Lulwah M. Alkwai, Michael L. Nelson, Michele C. Weigle (a) Response to the request http://tripadvisor.com/where_to_travel at the Internet Archive (b) Proposed recommendations for the requested URI http://tripadvisor.com/where_to_ travel displayed with MementoEmbed [15] social cards Figure 1: The actual response to the requested URI http://tripadvisor.com/where_to_travel (1a) and its proposed replacement (1b) and used those as queries to the web archive. They measured the Klein et al. [20] addressed a similar but slightly different problem coverage of the archived content retrieved by the current search by using web archives to recommend replacement pages on the engine and found that on page one of Bing results, 94% are available live web. They investigated four techniques for using the archived in the Internet Archive. Note that this technique will not find URLs page to generate queries for live web search engines: (1) lexical that have been missing (HTTP status 404) long enough for Bing to signatures, (2) web page titles, (3) tags, and (4) link neighborhood have removed them from its index. lexical signatures. Using these four methods helped to find a replace- ment for missing web pages. Various datasets were used, including Making Recommendations from Web Archives for “Lost” Web Pages DMOZ. By comparing the different methods, they found that 70% Rajalakshmi et al. [25] proposed an approach where N-gram of the web pages were recovered using the title method. The result based features are extracted from URIs alone, and the URI is classi- increased to 77% by combining the other three methods. In their fied using Support Vector Machines and Maximum Entropy Classi- work, the user will get a single alternative when a page is not found fiers. In this work, they used the 3-gram features from the URIon on the live Web. two datasets: 2 million URIs from DMOZ and a WebKB dataset with Huurdeman et al. [13, 14] detailed their approach to recover 4K URIs. Using this method on the WebKB dataset resulted in an in- pages in the unarchived Web based on the existence of links and crease of F1 score by 20.5% compared to the related work [11, 17, 18]. anchors of crawled pages. The data used was from the Dutch 2012 Also, using this method on DMOZ resulted in an increase of F1 National Library of the Netherlands2 (KB). Both external links (inter- score by 4.7% compared to the related work [8, 18, 24]. server links), which are links between different servers, and site One of the features we will use to rank the candidate URIs is internal links (intra-server links), which occur within a server, were the archival quality. Archival quality refers to measuring memento included in the dataset. Their findings included that the archived damage by evaluating the impact of missing resources in a web pages show evidence of a large number of unarchived pages and page. The missing resources could be text, images, video, audio, web sites. Finally, they found that even with a few words to describe style sheet, or any other type of resource on the web page. Brunelle a missing web page, they can be found within the first rank. et al. [10] proposed a damage rating algorithm to measure the Classification is the process of comparing representations ofdoc- relative value of embedded resources and evaluate archival success. uments with representations of labeled categories and computing The algorithm is based on a URI’s MIME type, size, and location similarity to find to which category the documents belong. Baykan of the embedded resources. In the Internet Archive the average et al.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    12 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us