Leveraging Content from Open Corpus Sources for Technology Enhanced Learning

Total Page:16

File Type:pdf, Size:1020Kb

Leveraging Content from Open Corpus Sources for Technology Enhanced Learning Leveraging Content from Open Corpus Sources for Technology Enhanced Learning A Thesis submitted to the University of Dublin, Trinity College for the degree of Doctor in Philosophy Séamus Lawless Knowledge and Data Engineering Group, School of Computer Science and Statistics, Trinity College Dublin Submitted October 2009 Declaration I, the undersigned, declare that this work has not been previously submitted as an exercise for a degree at this or any other University, and that, unless otherwise stated, it is entirely my own work. ________________________ Séamus Lawless October 2009 ii Permission to lend or copy I, the undersigned, agree that the Trinity College Library may lend or copy this thesis upon request. ________________________ Séamus Lawless October 2009 iii ACKNOWLEGDEMENTS Many people have influenced the writing of this thesis and provided both guidance and support on what has been a challenging, yet enjoyable and rewarding journey. I would like to acknowledge these individuals and thank them for the contributions they have made over the past number years which have made this thesis and the work described herein possible. Firstly, I would like to thank my supervisor, Prof. Vincent Wade, whose vast knowledge and experience, patience and encouragement has made this work possible. Special thanks are also reserved for Dr. Lucy Hederman, my co-supervisor during much of this research, whose insightful input was invaluable. As importantly, I would like to thank my family. My parents, Betty and Jimmy, for their unconditional and unfailing love, encouragement, belief and guidance throughout my life. They have set an example which I will always aspire to live up to. My fiancée, Pam, whose love, kindness, encouragement and patience helped me persevere on what has been a long and sometimes difficult road. My brother and sisters, for always being there, whenever and wherever I have needed them. My family’s support and boundless humour have been the foundation for everything I have achieved. I would also like to extend my gratitude to the members of the Knowledge and Data Engineering Group (KDEG) in Trinity College Dublin. Their friendship and insightful contributions have had a significant impact on this research. A number of us have undertaken this journey together and have always provided a mutual willing ear during the difficult periods. Thanks are also due to my colleagues in the School of Computer Science and Statistics. Finally, I would like to express my gratitude to the Irish Research Council for Science, Engineering and Technology for funding the research detailed in this thesis. I sincerely thank you all. Although nature commences with reason and ends in experience it is necessary for us to do the opposite, that is to commence with experience and from this to proceed to investigate the reason. – Leonardo Da Vinci (1452 – 1519) iv ABSTRACT As educators attempt to incorporate the use of educational technologies in course curricula, the lack of appropriate and accessible digital content resources acts as a barrier to adoption. Quality educational digital resources can prove expensive to develop and have traditionally been restricted to use in the environment in which they were authored. As a result, educators who wish to adopt these approaches are compelled to produce large quantities of high quality educational content. This can lead to excessive workloads being placed upon the educator, whose efforts are better exerted on the pedagogical aspects of eLearning design. The accessibility, portability, repurposing and reuse of digital resources thus became, and remain, major challenges. The key motivation of this research is to enable the utilisation of the vast amounts of accumulated knowledge and educational content accessible via the World Wide Web in Technology Enhanced Learning (TEL). This thesis proposes an innovative approach to the targeted sourcing of open corpus content from the WWW and the resource-level reuse of such content in pedagogically beneficial TEL offerings. The thesis describes the requirements, both educational and technical, for a tool-chain that enables the discovery, classification, harvesting and delivery of content from the WWW, and a novel TEL application which demonstrates the resource-level reuse of open corpus content in the execution of a pedagogically meaningful educational offering. Presented in this work are the theoretical foundations, design and implementation of two applications: the Open Corpus Content Service (OCCS); and the User-driven Content Retrieval, Exploration and Assembly Toolkit for eLearning (U-CREATe). To evaluate and validate this research, a detailed analysis of the different aspects of the research is presented, outlining and addressing the discovery, classification and harvesting of open corpus content from the WWW and open corpus content utilisation in TEL. This analysis provides confidence in the ability of the OCCS to generate collections of highly relevant open corpus content in defined subject areas. The analysis also provides confidence that the resource-level reuse of such content in educational offerings is possible, and that these educational offerings can be pedagogically beneficial to the learner. A novel approach to the sourcing of open corpus educational content for integration and reuse in TEL is the primary contribution to the State of the Art made by this thesis and the research described therein. This approach differs significantly from those used by current TEL systems in the creation of learning offerings and provides a service which is considerably different to that offered by general purpose web search engines. v TABLE OF CONTENTS ACKNOWLEGDEMENTS ..........................................................................................iv ABSTRACT ................................................................................................................ v TABLE OF CONTENTS .............................................................................................vi TABLE OF FIGURES ................................................................................................ xii ABBREVIATIONS .................................................................................................... xvi 1 Introduction ......................................................................................................... 1 1.1 Motivation ...................................................................................................... 1 1.2 Research Question ....................................................................................... 5 1.3 Research Goals and Objectives .................................................................... 6 1.4 Research Contribution .................................................................................. 8 1.5 Research Approach ...................................................................................... 9 2 Technology Enhanced Learning and Educational Content ............................... 11 2.1 Introduction to Technology Enhanced Learning .......................................... 11 2.2 Educational Theories in Technology Enhanced Learning ........................... 12 2.2.1 Theoretical Categorisations of Learning ............................................... 13 2.2.1.1 Associationist/Empiricist Perspective ............................................. 13 2.2.1.2 Cognitive Perspective .................................................................... 14 2.2.1.3 Situative Perspective ...................................................................... 16 2.2.2 Mapping Educational Theory to the Pedagogical Design of Learning Environments .................................................................................................... 17 2.2.3 Analysis ................................................................................................ 20 2.2.4 Summary .............................................................................................. 21 2.3 Educational Content Creation, Dissemination and Utilisation...................... 21 2.3.1 Learning Objects and Content Modelling Standards ............................ 22 2.3.1.1 Analysis .......................................................................................... 25 2.3.2 Digital Content Repositories ................................................................. 27 2.3.2.1 Analysis .......................................................................................... 30 2.3.3 Content Publication, Aggregation and Social Applications ................... 34 2.3.3.1 Analysis .......................................................................................... 38 2.3.4 Content Utilisation in Personalised TEL................................................ 39 2.3.4.1 Analysis .......................................................................................... 42 vi 2.3.5 Digital Rights Management and Intellectual Property ........................... 42 2.3.6 Summary .............................................................................................. 43 2.4 Summary ..................................................................................................... 44 3 State of the Art of Information Retrieval on the WWW ...................................... 45 3.1 Introduction ................................................................................................. 45 3.2 The Evolution of Information Retrieval .......................................................
Recommended publications
  • Crawling Frontier Controls
    Nutch – ApacheCon US '09 Web-scale search engine toolkit search Web-scale Today and tomorrow Today Apache Andrzej Białecki [email protected] Nutch – ApacheCon US '09 • • • Questions answers and future Nutch present and solutions)some (and Challenges Nutchworkflow: Nutcharchitecture overview general Web in crawling project the About Searching Crawling Setup Agenda 2 Nutch – ApacheCon US '09 • • Collections typically 1 mln - 200 mln documents mln Collections -typically 200 1 mln search mostly vertical in operation, installations Many Spin-offs: (sub-project Lucene) of Apache project since 2004 Mike Cafarella creator, and Lucene bythe Cutting, Doug 2003 in Founded Content type detection and parsing Tika → Map-Reduce and distributed → Hadoop FS Apache Nutch project 3 Nutch – ApacheCon US '09 4 Nutch – ApacheCon US '09 first, random Traversal: depth- breadth-first, edges, the follow listsas Oftenadjacency represented (neighbor) <alabels: href=”..”>anchor Edge text</a> Edges (links): hyperlinks like <a href=”targetUrl”/> Nodes (vertices):URL-s identifiers as unique 6 2 8 1 3 Web as a directed graph 5 4 7 9 7 →3, 4, 8, 9 5 →6, 9 1 →2, 3, 4, 5, 6 5 Nutch – ApacheCon US '09 … What's in a search engine? a fewa things may surprisethat you! 6 Nutch – ApacheCon US '09 Injector -links(in/out) - Web graph pageinfo Search engine building blocks Scheduler Updater Crawling frontierCrawling controls Crawler repository Content Searcher Indexer Parser 7 Nutch – ApacheCon US '09 Robust API and integration options Robust APIintegration and Full-text&indexer search engine processingdata framework Scalable Robustcontrols frontier crawling processing (parsing, content filtering) Plugin-based crawler distributed multi-threaded, Multi-protocol, modular: highly Plugin-based, graph) (web link database and database Page − − − − Support Support for search distributed or Using Lucene Solr Map-reduce processing Mostvia plugins be behavior can changed Nutch features at a glance 8 Hadoop foundation File system abstraction • Local FS, or • Distributed FS − also Amazon S3, Kosmos and other FS impl.
    [Show full text]
  • The Google Search Engine
    University of Business and Technology in Kosovo UBT Knowledge Center Theses and Dissertations Student Work Summer 6-2010 The Google search engine Ganimete Perçuku Follow this and additional works at: https://knowledgecenter.ubt-uni.net/etd Part of the Computer Sciences Commons Faculty of Computer Sciences and Engineering The Google search engine (Bachelor Degree) Ganimete Perçuku – Hasani June, 2010 Prishtinë Faculty of Computer Sciences and Engineering Bachelor Degree Academic Year 2008 – 2009 Student: Ganimete Perçuku – Hasani The Google search engine Supervisor: Dr. Bekim Gashi 09/06/2010 This thesis is submitted in partial fulfillment of the requirements for a Bachelor Degree Abstrakt Përgjithësisht makina kërkuese Google paraqitet si sistemi i kompjuterëve të projektuar për kërkimin e informatave në ueb. Google mundohet t’i kuptojë kërkesat e njerëzve në mënyrë “njerëzore”, dhe t’iu kthej atyre përgjigjen në formën të qartë. Por, ky synim nuk është as afër ideales dhe realizimi i tij sa vjen e vështirësohet me zgjerimin eksponencial që sot po përjeton ueb-i. Google, paraqitet duke ngërthyer në vetvete shqyrtimin e pjesëve që e përbëjnë, atyre në të cilat sistemi mbështetet, dhe rrethinave tjera që i mundësojnë sistemit të funksionojë pa probleme apo të përtërihet lehtë nga ndonjë dështim eventual. Procesi i grumbullimit të të dhënave ne Google dhe paraqitja e tyre në rezultatet e kërkimit ngërthen në vete regjistrimin e të dhënave nga ueb-faqe të ndryshme dhe vendosjen e tyre në rezervuarin e sistemit, përkatësisht në bazën e të dhënave ku edhe realizohen pyetësorët që kthejnë rezultatet e radhitura në mënyrën e caktuar nga algoritmi i Google.
    [Show full text]
  • Web Crawling, Analysis and Archiving
    Web Crawling, Analysis and Archiving Vangelis Banos Aristotle University of Thessaloniki Faculty of Sciences School of Informatics Doctoral dissertation under the supervision of Professor Yannis Manolopoulos October 2015 Ανάκτηση, Ανάλυση και Αρχειοθέτηση του Παγκόσμιου Ιστού Ευάγγελος Μπάνος Αριστοτέλειο Πανεπιστήμιο Θεσσαλονίκης Σχολή Θετικών Επιστημών Τμήμα Πληροφορικής Διδακτορική Διατριβή υπό την επίβλεψη του Καθηγητή Ιωάννη Μανωλόπουλου Οκτώβριος 2015 i Web Crawling, Analysis and Archiving PhD Dissertation ©Copyright by Vangelis Banos, 2015. All rights reserved. The Doctoral Dissertation was submitted to the the School of Informatics, Faculty of Sci- ences, Aristotle University of Thessaloniki. Defence Date: 30/10/2015. Examination Committee Yannis Manolopoulos, Professor, Department of Informatics, Aristotle University of Thes- saloniki, Greece. Supervisor Apostolos Papadopoulos, Assistant Professor, Department of Informatics, Aristotle Univer- sity of Thessaloniki, Greece. Advisory Committee Member Dimitrios Katsaros, Assistant Professor, Department of Electrical & Computer Engineering, University of Thessaly, Volos, Greece. Advisory Committee Member Athena Vakali, Professor, Department of Informatics, Aristotle University of Thessaloniki, Greece. Anastasios Gounaris, Assistant Professor, Department of Informatics, Aristotle University of Thessaloniki, Greece. Georgios Evangelidis, Professor, Department of Applied Informatics, University of Mace- donia, Greece. Sarantos Kapidakis, Professor, Department of Archives, Library Science and Museology, Ionian University, Greece. Abstract The Web is increasingly important for all aspects of our society, culture and economy. Web archiving is the process of gathering digital materials from the Web, ingesting it, ensuring that these materials are preserved in an archive, and making the collected materials available for future use and research. Web archiving is a difficult problem due to organizational and technical reasons. We focus on the technical aspects of Web archiving.
    [Show full text]
  • Detecting Malicious Websites with Low-Interaction Honeyclients
    Monkey-Spider: Detecting Malicious Websites with Low-Interaction Honeyclients Ali Ikinci Thorsten Holz Felix Freiling University of Mannheim Mannheim, Germany Abstract: Client-side attacks are on the rise: malicious websites that exploit vulner- abilities in the visitor’s browser are posing a serious threat to client security, compro- mising innocent users who visit these sites without having a patched web browser. Currently, there is neither a freely available comprehensive database of threats on the Web nor sufficient freely available tools to build such a database. In this work, we in- troduce the Monkey-Spider project [Mon]. Utilizing it as a client honeypot, we portray the challenge in such an approach and evaluate our system as a high-speed, Internet- scale analysis tool to build a database of threats found in the wild. Furthermore, we evaluate the system by analyzing different crawls performed during a period of three months and present the lessons learned. 1 Introduction The Internet is growing and evolving every day. More and more people are becoming part of the so-called Internet community. With this growth, also the amount of threats for these people is increasing. Online criminals who want to destroy, cheat, con others, or steal goods are evolving rapidly [Ver03]. Currently, there is no comprehensive and free database to study malicious websites found on the Internet. Malicious websites are websites which have any kind of content that could be a threat for the security of the clients requesting these sites. For example, a malicious website could exploit a vulnerability in the visitor’s web browser and use this to compromise the system and install malware on it.
    [Show full text]
  • Web Manifestations of Knowledge-Based Innovation Systems in the U.K
    Web Manifestations of Knowledge-based Innovation Systems in the U.K. David Patrick Stuart B.A.(hons) A thesis submitted in partial fulfilment of the requirements of the University of Wolverhampton for the degree of Doctor of Philosophy January 2008 This work or any part thereof has not previously been presented in any form to the University or to any other body whether for the purposes of assessment, publication or for any other purpose (unless otherwise indicated). Save for any express acknowledgments, references and/or bibliographies cited in the work, I confirm that the intellectual content of the work is the result of my own efforts and of no other person. The right of David Stuart to be identified as author of this work is asserted in accordance with ss.77 and 78 of the Copyright, Designs and Patents Act 1988. At this date copyright is owned by the author. Signature……………………………………….. Date…………………………………………….. Publication List Journal Papers: Stuart, D., & Thelwall, M. (2006). Investigating triple helix relationships using URL citations: a case study of the UK West Midlands automobile industry. Research Evaluation, 15(2), 97-106. Stuart, D., Thelwall, M., & Harries, G. (2007). UK academic web links and collaboration – an exploratory study. Journal of Information Science, 33(2), 231-246. Thelwall, M., & Stuart, D. (2006). Web crawling ethics revisited: Cost, privacy, and denial of service. Journal of the American Society for Information Science and Technology, 57(13), 1771-1779. Conference Papers: Stuart, D., & Thelwall, M. (2005). What can university-to-government web links reveal about university-government collaborations? In P. Ingwersen, & B.
    [Show full text]
  • Data-Intensive Text Processing with Mapreduce
    Data-Intensive Text Processing with MapReduce Synthesis Lectures on Human Language Technologies Editor Graeme Hirst, University of Toronto Synthesis Lectures on Human Language Technologies is edited by Graeme Hirst of the University of Toronto.The series consists of 50- to 150-page monographs on topics relating to natural language processing, computational linguistics, information retrieval, and spoken language understanding. Emphasis is on important new techniques, on new applications, and on topics that combine two or more HLT subfields. Data-Intensive Text Processing with MapReduce Jimmy Lin and Chris Dyer 2010 Semantic Role Labeling Martha Palmer, Daniel Gildea, and Nianwen Xue 2010 Spoken Dialogue Systems Kristiina Jokinen and Michael McTear 2009 Introduction to Chinese Natural Language Processing Kam-Fai Wong, Wenjie Li, Ruifeng Xu, and Zheng-sheng Zhang 2009 Introduction to Linguistic Annotation and Text Analytics Graham Wilcock 2009 Dependency Parsing Sandra Kübler, Ryan McDonald, and Joakim Nivre 2009 Statistical Language Models for Information Retrieval ChengXiang Zhai 2008 Copyright © 2010 by Morgan & Claypool All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means—electronic, mechanical, photocopy, recording, or any other except for brief quotations in printed reviews, without the prior permission of the publisher. Data-Intensive Text Processing with MapReduce Jimmy Lin and Chris Dyer www.morganclaypool.com ISBN: 9781608453429 paperback ISBN:
    [Show full text]
  • Bachelor of Library and Information Science (Blisc.)
    Bachelor of Library and Information Science (B.L.I.Sc.) Goa University Goa University Library, Taleigao Plateau Goa PIN 403 206 Website: www.unigoa.ac.in Tel: 0832-6519072/272/087, Fax: 0832-2451184 Bachelor of Library and Information Science (B.L.I.Sc.) Programme Duration of the Programme : One Year Programme (Credit based two semesters) Objectives of the Programme : • To train students for a professional career in Library and Information Services • To produce a quality manpower for collection, compilation and dissemination of information products and services in and beyond conventional libraries and information centres Number of seats : 25 Availability and Reservation of Seats Total number seats available for admission are 25. University shall allocate seats as per State Government/University policy. Accordingly, the distribution of seats will be as follows. Sl No Category Seats 1 OBC 7 2 SC 1 3 ST 3 4 Differentially abled 1 5 General 11 6 Other Universities 2 Total 25 Qualification for admission : Graduates in any faculty like Languages and Literature, Social sciences, Commerce, Natural Sciences, Life Sciences and Environment, Engineering, Medicine, etc with minimum 40% aggregate marks at their graduation from any recognized university in India or abroad are eligible to apply for the B.L.I.Sc. Evaluation : The assessment of the Programmes comprises of continuous Intra-Semester Assessment (ISA) and Semester End Assessment (SEA) and is done fully internally. Fee details : For Outside Item For Goa University Students Students Tuition Fee 18900 18900 Reg. Fee (GU Students) 500 0 (Outside Students) 0 2300 Gym, Stud. Union, ID Card 410 410 Student Aid Fund 120 120 Computer Lab facility 810 810 Annual Internet Fee 230 230 Annual Library Fee 460 460 Caution Deposit (Refundable) 1750 1750 Total 23180 24980 Curriculum details : Curriculum details for B.L.I.Sc.
    [Show full text]
  • A Focused Web Crawler Using Link and Content Analysis for Relevence Prediction
    PDD CRAWLER: A FOCUSED WEB CRAWLER USING LINK AND CONTENT ANALYSIS FOR RELEVENCE PREDICTION. Prashant Dahiwale1, M M Raghuwanshi2 and Latesh Malik3 1Research Scholar , Dept of CSE, GHRCE and Assist Prof Dept of CSE, RGCER,Nagpur, India [email protected] 2Department of Computer Science Engineering, RGCER, Nagpur India [email protected] 3Department of Computer science Engineering, GHRCE, Nagpur, India [email protected] ABSTRACT Majority of the computer or mobile phone enthusiasts make use of the web for searching activity. Web search engines are used for the searching; The results that the search engines get are provided to it by a software module known as the Web Crawler. The size of this web is increasing round-the-clock. The principal problem is to search this huge database for specific information. To state whether a web page is relevant to a search topic is a dilemma. This paper proposes a crawler called as “PDD crawler” which will follow both a link based as well as a content based approach. This crawler follows a completely new crawling strategy to compute the relevance of the page. It analyses the content of the page based on the information contained in various tags within the HTML source code and then computes the total weight of the page. The page with the highest weight, thus has the maximum content and highest relevance. KEYWORDS Web Crawler, HTML, Tags, Searching Relevance, Metadata 1. INTRODUCTION The World Wide Web is a huge collection of data. The data keeps increasing week by week, day by day and hour by hour.
    [Show full text]
  • Enhancing Web Search User Experience: from Document Retrieval to Task Recommendation
    Universita` Ca' Foscari di Venezia Dottorato di Ricerca in Informatica Scuola di Dottorato in Scienze e Tecnologie - XXIII Ciclo (A.A. 2010{2011) Tesi di Dottorato Settore Scientifico Disciplinare: INF/01 Enhancing Web Search User Experience: from Document Retrieval to Task Recommendation Gabriele Tolomei matricola n. 955515 Tutore del Dottorando Prof. Salvatore Orlando Direttore della Scuola Prof. Paolo Ugo Ottobre 2011 Pagina Web dell'autore: http://miles.isti.cnr.it/~tolomei Posta elettronica dell'autore: [email protected] Indirizzo dell'autore: Dipartimento di Informatica Universit`aCa' Foscari di Venezia Via Torino, 155 30172 Venezia Mestre { Italia tel. +39 041 2348411 fax. +39 041 2348419 web: http://www.dsi.unive.it A Nonna Giordana Abstract The World Wide Web (i.e., Web) is the biggest and most heterogeneous database that humans have ever built, making it the place of choice where people look at whenever they come up with any sort of information need. Today, most of Web users put their trust in Web search engines for pursuing and satisfying their infor- mation needs. Of course, modern Web search engines radically evolved since their first appear- ance almost fifteen years ago. Indeed, nowadays they provide a still growing set of capabilities for enhancing users' experience during the whole search process. Nev- ertheless, they are still in essence Web documents retrieval tools, namely Web-scale information retrieval systems, whereas users' expectations and needs are increas- ingly becoming more complex and heterogeneous. This trend is actually confirmed by a growing \addiction to Web search": no matter what an information need is, user is anyway brought to ask for it to a Web search engine, which will hopefully give the answer she expects.
    [Show full text]
  • Department of Library and Information Science Svu College of Arts Tirupati
    DEPARTMENT OF LIBRARY AND INFORMATION SCIENCE SVU COLLEGE OF ARTS TIRUPATI Program Outcomes (POs): It represent the knowledge, skills and attitudes the students should have at the end of a course completion of their respective Library and Information Science program. Program Specific Outcomes (PSOs): These are statements that defines outcomes of a program which makes students realize the fact that the knowledge and techniques learnt in this course has direct implication for the betterment of society and its sustainability Course Outcomes (COs): It gives the resultant knowledge and skills the student acquires at the end of each course. It defines the cognitive processes a course provides. PROGRAM OUTCOMES Preparing the learners to acquire professional Skills, Information and PO 1 Communication Technology Skills to get Job opportunities as teaching faculty in Library and Information Science departments and as Librarians in different types of libraries. PO 2 Making the learners to identify, formulate, review research literature and analyse different problems to reach conclusions using principles of library management PO 3 Enabling the learners to design solutions for complex library problems in order to satisfy the various approaches of the users Motivating the learners to conduct investigations of multifaceted problems PO 4 by applying research-based knowledge and different types of research methods including conducting of user studies and case studies in libraries, analysis and interpretation of data and synthesis of the information to
    [Show full text]
  • Data-Intensive Text Processing with Mapreduce
    i Data-Intensive Text Processing with MapReduce Jimmy Lin and Chris Dyer University of Maryland, College Park Manuscript prepared April 11, 2010 This is the pre-production manuscript of a book in the Morgan & Claypool Synthesis Lectures on Human Language Technologies. Anticipated publication date is mid-2010. ii Contents Contents.....................................................................ii 1 Introduction . 1 1.1 Computing in the Clouds. .6 1.2 Big Ideas. .9 1.3 Why Is This Different? . 15 1.4 What This Book Is Not . 17 2 MapReduce Basics . 18 2.1 Functional Programming Roots . 20 2.2 Mappers and Reducers . 22 2.3 The Execution Framework . 26 2.4 Partitioners and Combiners . 28 2.5 The Distributed File System . 31 2.6 Hadoop Cluster Architecture . 36 2.7 Summary . 38 3 MapReduce Algorithm Design. .39 3.1 Local Aggregation . 41 3.1.1 Combiners and In-Mapper Combining 41 3.1.2 Algorithmic Correctness with Local Aggregation 46 3.2 Pairs and Stripes. .50 3.3 Computing Relative Frequencies . 57 3.4 Secondary Sorting. .60 3.5 Relational Joins. .62 3.5.1 Reduce-Side Join 64 3.5.2 Map-Side Join 66 3.5.3 Memory-Backed Join 67 CONTENTS iii 3.6 Summary . 68 4 Inverted Indexing for Text Retrieval. .70 4.1 Web Crawling. .71 4.2 Inverted Indexes . 73 4.3 Inverted Indexing: Baseline Implementation . 75 4.4 Inverted Indexing: Revised Implementation. .77 4.5 Index Compression . 79 4.5.1 Byte-Aligned and Word-Aligned Codes 80 4.5.2 Bit-Aligned Codes 82 4.5.3 Postings Compression 84 4.6 What About Retrieval? .
    [Show full text]
  • Spiders, Crawlers, Harvesters, Bots
    Spiders, crawlers, harvesters, bots Thanks to B. Arms R. Mooney P. Baldi P. Frasconi P. Smyth C. Manning Last time • Evaluation of IR/Search systems – Quality of evaluation – Relevance – Evaluation is empirical – Measurements of Evaluation • Precision vs recall • F measure – Test Collections/TREC This time • Web crawlers • Crawler policy • Robots.txt • Scrapy Evaluation Query Engine Index Interface Indexer Users Crawler Web A Typical Web Search Engine Query Engine Index Interface Indexer Users Crawler Web A Typical Web Search Engine What is a Web Crawler? • The Web crawler is a foundational species! • Without crawlers, search engines would not exist. • But they get little credit! • Outline: What is a crawler How they work How they are controlled Robots.txt Issues of performance Research What a web crawler does • Gets data!!! • Can get fresh data. • Gets data for search engines: • Creates and repopulates search engines data by navigating the web, downloading documents and files • Follows hyperlinks from a crawl list and hyperlinks in the list • Without a crawler, there would be nothing to search Web crawler policies • The behavior of a Web crawler is the outcome of a combination of policies: • a selection policy that states which pages to download, • a re-visit policy that states when to check for changes to the pages, • a duplication policy • a politeness policy that states how to avoid overloading Web sites, and • a parallelization policy that states how to coordinate distributed Web crawlers. Crawlers vs Browsers vs Scrapers • Crawlers
    [Show full text]