Internals of an Aggregated Web News Feed

Total Page:16

File Type:pdf, Size:1020Kb

Internals of an Aggregated Web News Feed INTERNALS OF AN AGGREGATED WEB NEWS FEED Mitja Trampuš, Blaž Novak Artificial Intelligence Laboratory Jozef Stefan Institute Jamova 39, 1000 Ljubljana, Slovenia e-mail: {mitja.trampus, blaz.novak}@ijs.si ABSTRACT 3) Parses each article to obtain We present IJS NewsFeed, a pipeline for acquiring a a. Potential new RSS sources, to be used in clean, continuous, real-time aggregated stream of step (1) publically available news articles from web sites across b. Cleartext version of the article body the world. 4) Process articles with Enrycher (see Section 3.2) The articles are stripped of the web page chrome and 5) Expose two streams of news articles (cleartext and semantically enriched to include e.g. a list of entities Enrycher-processed) to end users. appearing in each article. The results are cached and distributed in an efficient manner. 2 SYSTEM ARCHITECTURE Figure 1 gives a schematic overview of the architecture. The 1 INTRODUCTION first part of the aggregator is based around a PostgreSQL The news aggregator is a piece of software which provides database running on a Linux server. The database contains a a real-time aggregated stream of textual news items list of RSS feeds which are periodically downloaded by the provided by RSS-enabled news providers across the world. RSS monitoring component. RSS feeds contain a list of The pipeline performs the following main steps: news article URLs and some associated metadata, such as 1) Periodically crawls a list of RSS feeds and a subset tags, publication date, etc. Articles that are not already of Google News and obtains links to news articles present in the database are added to a list of article URLs, 2) Downloads the articles, taking care not to overload and marked for download. Tags and publication date are also any of the hosting servers stored alongside, if found in the RSS. Figure 1: The system architecture of IJS NewsFeed. A separate component periodically retrieves the list of new We initially implemented the algorithm by Pasternack articles and fetches them from the web. The complete because of its simplicity and reported state-of-the-art HTML is stored in the database, and simultaneously sent to performance. The algorithm scores each token (a word or a a set of cleaning processes over a 0mq message queue. tag) in the document based on how probable it is to comprise The cleaning process converts the HTML into UTF-8 the final result (the scores are trained); then it extracts the encoding, determines which part of the HTML contains the maximum token subsequence. useful text, and discards the remainder and all of the tags. Datasets Finally, a classifier is used to determine the primary We tested the initial algorithm on three manually developed language. datasets. Each of the three consists of 50 articles, each from The cleaned version of the text is stored back in the a different web site. database, and sent over a message queue to consumers. english – English articles only. Documents in English language are sent to the Enrycher web alphabet – Non-English articles using an alphabet, i.e. service, where named entities are extracted and resolved, one glyph per sound. This includes e.g. Arabic. and the entire document is categorized into a DMOZ topic hierarchy. syllabary – Non-English articles using a syllabary, i.e. one glyph per syllable. This boils down to Asian Both the cleartext and the enriched versions of documents languages. They lack word boundaries and have are fed to a filesystem cache, which stores a sequence of generally shorter articles in terms of glyphs. Also, the compressed xml files, each containing a series of documents design of Asian pages tends to be slightly different. in the order they have arrived through the processing pipeline. The caching service exposes an HTTP interface to Some of the input pages (about 5%), realistically, also do not the world through an Apache transparent proxy, serving include meaningful content. This is different from other data those compressed xml files on user request. sets but very relevant to our scenario. Examples are paywall pages and pages with a picture bearing a single-sentence The Apache server also hosts a CGI process capable of caption. generating HTML5 server-side events, which contains the article metadata and cleartext as payload. These events can The fact that each of the 150 articles comes from a different be consumed using Javascripts EventSource object in a web site is crucial – most of the papers on this topic evaluate on a browser. dataset from a small number of sites, which leads to overfitting and poor performance in the general case. This was also the case with Pasternack’s algorithm. As the 3 DATA PREPROCESSING performance was unsatisfactory, we developed three new Data preprocessing is an important part of the pipeline, both algorithms. in terms of the added value provides and in terms of Algorithms challenges posed by the data volume. The articles WWW – an improved version of Pasternack (2009), it themselves are certainly useful, but almost any automated extracts two most promising contiguous chunks of text task dealing with them first needs to transform the raw from the article to account for the fact that the first HTML into a form more suitable for further processing. We paragraph is often placed separately from the main therefore perform the preprocessing ourselves; this is much article body. like the practice followed by professional data aggregation services like Spinn3r or Gnip. WWW++ – a combination of WWW and heuristic pre- and post-processing to account for the most obvious In terms of data volume, preprocessing is the most errors of WWW. For instance, preprocessing tries to interesting stage and the one at which the most tradeoff can remove user comments. be made. The present data download rate of about one article per second is nothing extreme, especially if we consider DOM – a completely heuristics-based approach scaling to multiple processing nodes; however, it is proposed here which requires the DOM tree to be nontrivial in that adding complex preprocessing steps (e.g. computed. With the fast libxml package, this is not a full syntactic parsing of text) or drastically increasing data limiting factor. The core of the heuristic is to take the load (e.g. including a 10% sample of the Twitter feed) first large enough DOM element that contains enough would turn preprocessing into a bottleneck and require us to promising <p> elements. Failing that, take the first <td> scale the architecture. or <div> element which contains enough promising text. The heuristics for the definition of “promising” 3.1 Extracting article body from web pages rely on metrics found in other papers as well; most Extracting meaningful content from the HTML is the most importantly, the amount of markup within a node. obviously needed preprocessing step. As this is a pervasive Importantly, none of the heuristics are site-specific. problem, a lot has been published on the topic; see e.g. Pasternack (2009), Arias (2009), and Kohlschütter (2010). Number of cleartext articles 120000 100000 80000 60000 40000 20000 0 2008-05-06 2008-05-19 2008-06-01 2008-06-14 2008-06-27 2008-07-11 2008-07-24 2008-08-06 2008-08-19 2008-09-01 2008-09-14 2008-09-27 2008-10-18 2008-11-15 2008-11-30 2008-12-13 2008-12-31 2009-01-13 2009-01-26 2009-02-08 2009-02-21 2009-03-06 2009-03-19 2009-04-01 2009-04-14 2009-04-27 2009-05-18 2009-06-10 2009-06-23 2009-07-06 2009-07-19 2009-08-01 2009-08-29 2010-11-02 2010-11-15 2010-11-28 2010-12-11 2010-12-24 2011-01-06 2011-01-19 2011-02-01 2011-02-14 2011-02-27 2011-03-12 2011-03-25 2011-04-07 2011-04-20 2011-05-03 2011-05-16 2011-06-25 2011-07-17 2011-08-26 2011-09-08 2011-09-21 2011-10-04 2011-10-17 2011-10-30 2012-01-16 2012-01-29 2012-02-11 2012-02-24 2012-03-08 2012-03-21 2012-04-03 2012-04-16 2012-04-29 2012-05-12 2012-05-25 2012-06-07 2012-06-20 2012-07-03 2012-07-16 2012-07-29 2012-08-11 Figure 2: The daily number of downloaded articles. A weekly pattern is nicely observable. Through most of 2011, only Google News was used as an article source, hence the significantly lower volume in that period. In all three algorithms, all pages are first normalized to the Garbage [5.8%] – The output contains mostly or UTF-8 character set using the BeautifulSoup package exclusively text that is not in the golden standard. These (which in turn uses a combination of http headers, meta tags are almost always articles with a very short body and a and the chardet tool). long copyright disclaimer that gets picked up instead. Evaluation Missed [5.8%] – Although the article contains We evaluated two of the three pairs of algorithms by meaningful content, the output is an empty string, i.e. comparing per-article performance. We did compare WWW the algorithm fails to find any content. and DOM; based on informal inspection of outputs, DOM If we combine “Perfect” and “Good” (where the outcome is would be certain to perform better. most often only a sentence away from the perfect match) Algo WWW vs WWW++ WWW++ vs DOM into a “Positive” score, both precision and recall for DOM number of articles where one number of articles where one are 94%.
Recommended publications
  • Specifications for Implementing Web Feeds in DLXS Kevin S
    Hawkins 1/5/2011 5:01:52 PM Page 1 of 5 * SCHOLARLY PUBLISHING OFFICE WHITEPAPER Specifications for implementing web feeds in DLXS Kevin S. Hawkins Executive Summary SPO uses DLXS to deliver nearly all of its publications online, but this software does not offer any functionality for web feeds (commonly called RSS feeds). The components of web feeds are reviewed, and recommendations are made for implementing Atom 1.0 web feeds in SPO publications using a script run during releasing which determines which content to push to the feed. Motivations Members of the editorial board of the Journal of Electronic Publishing (collid jep) and SPO staff have requested that web feed functionality be provided for SPO publications. This would allow end users to subscribe to feeds of not only serial content (the obvious use) but in theory any other collection to which content is added, especially on a regular basis. A feed could provide a single notice that a collection has been updated, perhaps with a table of contents (henceforth a small feed), or instead contain separate notices for each item added to the collection (a large feed). Components of a web feed A web feed operates with three necessary components: • a website that syndicates (or publishes) a feed • a feed standard being used by the website • the end user’s news aggregator (feed reader or news reader), which retrieves the feed periodically. A feed is an XML document containing individual entries (also items, stories, or articles). Syndicating a feed The content provider’s website often has buttons (icons called chicklets) to be used by the end user to add a feed to his or her news aggregator.
    [Show full text]
  • EMERGING TECHNOLOGIES Skype and Podcasting: Disruptive Technologies for Language Learning
    Language Learning & Technology September 2005, Volume 9, Number 3 http://llt.msu.edu/vol9num3/emerging/ pp. 9-12 EMERGING TECHNOLOGIES Skype and Podcasting: Disruptive Technologies for Language Learning Robert Godwin-Jones Virginia Comonwealth University New technologies, or new uses of existing technologies, continue to provide unique opportunities for language learning. This is in particular the case for several new network options for oral language practice. Both Skype and podcasting can be considered "disruptive technologies" in that they allow for new and different ways of doing familiar tasks, and in the process, may threaten traditional industries. Skype, the "people's telephone," is a free, Internet-based alternative to commercial phone service, while podcasting, the "radio for the people," provides a "narrowcasting" version of broadcast media. Both have sparked intense interest and have large numbers of users, although it is too soon to toll the bell for telephone companies and the radio industry. Skype and podcasting have had a political aspect to their embrace by early adopters -- a way of democratizing institutions -- but as they reach the mainstream, that is likely to become less important than the low cost and convenience the technologies offer. Both technologies offer intriguing opportunities for language professionals and learners, as they provide additional channels for oral communication. Skype and Internet Telephony Skype is a software product which provides telephone service through VoIP (Voice over IP), allowing your personal computer to act like a telephone. A microphone attached to the computer is necessary and headphones are desirable (to prevent echoes of the voice of your conversation partner). It is not the only such tool, nor the first, but because it provides good quality (through highly efficient compression) and is free, it has become widely used.
    [Show full text]
  • Web 2.0: Beyond the Concept Practical Ways to Implement RSS, Podcasts, and Wikis
    Web 2.0: Beyond the Concept Practical Ways to Implement RSS, Podcasts, and Wikis By Karen Huffman, National Geographic Society Abstract This article focuses on how the National Geographic Society’s Libraries & Information Services has implemented RSS, podcasts and wikis with suggested practical application for academic libraries. The article is an adaptation from a presentation given during the Special Libraries Association’s Annual Conference, June 2006 (Huffman & Ferry, 2006). The Sparks That Ignited Us…. National Geographic’s Libraries & Information Services (LIS) discovery and implementation of technologies dubbed “Web 2.0” has been an evolutionary, rather than revolutionary process. We watched the trends to evolve our products and services. In 2005 we began seeing online news sources like BBCnews.com offering alternative ways to access their content. New RSS-labeled icons appeared in the margins of websites with links to directories of feeds. These RSS directories generally organized their feeds either topically or by news section and often included frequently asked questions about RSS as well as recommended feed readers. As information organizers (and multi-taskers who streamline their work processes wherever possible), we initially saw RSS feeds as a way to keep current with news sources, blogs, and research sites that were regularly publishing content through XML-based feeds! We saw the arrival of podcasts on the heels of RSS. In January and February, 2005, the Pew Internet and American Life Project conducted a survey, and found 22 million, or 11%, Americans age 18 or older owned an iPod or MP3 player. Apple’s freely-available iTunes library brought podcasting to the mainstream with the launch of its “Podcasts” directory in June 2005.
    [Show full text]
  • Working with Feeds, RSS, and Atom
    CHAPTER 4 Working with Feeds, RSS, and Atom A fundamental enabling technology for mashups is syndication feeds, especially those packaged in XML. Feeds are documents used to transfer frequently updated digital content to users. This chapter introduces feeds, focusing on the specific examples of RSS and Atom. RSS and Atom are arguably the most widely used XML formats in the world. Indeed, there’s a good chance that any given web site provides some RSS or Atom feed—even if there is no XML-based API for the web site. Although RSS and Atom are the dominant feed format, other formats are also used to create feeds: JSON, PHP serialization, and CSV. I will also cover those formats in this chapter. So, why do feeds matter? Feeds give you structured information from applications that is easy to parse and reuse. Not only are feeds readily available, but there are many applications that use those feeds—all requiring no or very little programming effort from you. Indeed, there is an entire ecology of web feeds (the data formats, applications, producers, and consumers) that provides great potential for the remix and mashup of information—some of which is starting to be realized today. This chapter covers the following: * What feeds are and how they are used * The semantics and syntax of feeds, with a focus on RSS 2.0, RSS 1.0, and Atom 1.0 * The extension mechanism of RSS 2.0 and Atom 1.0 * How to get feeds from Flickr and other feed-producing applications and web sites * Feed formats other than RSS and Atom in the context of Flickr feeds * How feed autodiscovery can be used to find feeds * News aggregators for reading feeds and tools for validating and scraping feeds * How to remix and mashup feeds with Feedburner and Yahoo! Pipes Note In this chapter, I assume you have an understanding of the basics of XML, including XML namespaces and XML schemas.
    [Show full text]
  • Feed Your ‘Need to Read’ – an Explanation of RSS
    by Jen Sharp, jensharp.com Feed your ‘Need To Read’ – an explanation of RSS here exists a useful informational tool so powerful that provides feeds. For example, USA Today has separate that the government of China has completely feeds in its different sections. You know that a page has the banned its use since 2007. Yet many people have capabilities for RSS subscriptions by looking for the orange Tnot even heard about it, despite its juxtaposition in chicklet in the toolbar of your browser. Clicking on the logo front of our attention during our experiences on the web. will start you in the process of seeing what the feed for that Recently I was visiting with my dad, who is well-informed page looks like. Then you can choose how to subscribe. and somewhat technically savvy. I asked him if he had ever Choosing a reader can seem overwhelming, as there are heard of RSS feeds. He hadn’t, so when I pointed out the thousands of readers available, most of them for free. To ever-present orange “chicklet,” he remarked, “Oh! I thought start, decide how you what to view the information. RSS that was the logo for Home Depot!” feeds can be displayed in many ways, including: Very simply, RSS feeds are akin to picking up your ■ sent to your email inbox favorite newspaper and scanning the headlines for a story ■ as a personal browser start page like igoogle.com or you are interested in. Items are displayed with a synopsis and my.yahoo.com links to the full page of information.
    [Show full text]
  • Cloud Computing Bible
    Barrie Sosinsky Cloud Computing Bible Published by Wiley Publishing, Inc. 10475 Crosspoint Boulevard Indianapolis, IN 46256 www.wiley.com Copyright © 2011 by Wiley Publishing, Inc., Indianapolis, Indiana Published by Wiley Publishing, Inc., Indianapolis, Indiana Published simultaneously in Canada ISBN: 978-0-470-90356-8 Manufactured in the United States of America 10 9 8 7 6 5 4 3 2 1 No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 646-8600. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, 201-748-6011, fax 201-748-6008, or online at http://www.wiley.com/go/permissions. Limit of Liability/Disclaimer of Warranty: The publisher and the author make no representations or warranties with respect to the accuracy or completeness of the contents of this work and specifically disclaim all warranties, including without limitation warranties of fitness for a particular purpose. No warranty may be created or extended by sales or promotional materials. The advice and strategies contained herein may not be suitable for every situation. This work is sold with the understanding that the publisher is not engaged in rendering legal, accounting, or other professional services.
    [Show full text]
  • Content Syndication in News Aggregators. Towards Devaluation of Professional Journalistic Criterio
    Comunicar, n. 59, v. XXVII, 2019 | Media Education Research Journal | ISSN: 1134-3478; e-ISSN: 1988-3293 www.comunicarjournal.com Content syndication in news aggregators. Towards devaluation of professional journalistic criteria La sindicación de contenidos en los agregadores de noticias: Hacia la devaluación de los criterios profesionales periodísticos Dr. Concha Edo is Professor in the Department of Journalism and Global Communication at the Complutense University of Madrid (Spain) ([email protected]) (https://orcid.org/0000-0001-6960-867X) Dr. Juan Yunquera is Assistant Professor in the Department of Journalism and Audio-visual Communication at the University Carlos III of Madrid (Spain) ([email protected]) (https://orcid.org/0000-0002-4062-1659) Dr. Helder Bastos is Professor in the Department of Communication and Information Sciences at the University of Porto ([email protected]) (https://orcid.org/0000-0002-6351-4610) ABSTRACT The growing expansion of Internet access and mass-scale usage of social networking platforms and search engines have forced digital newspapers to deal with challenges, amongst which are the need to constantly update news, the increasing complexity of sources, the difficulty of exercising their function as gatekeepers in a fragmented environment in which the opinions, biases and preconceptions of pundits, their followers, Twitter users, etc. has taken on a new and decisive weight and the mounting pressure to publish certain news items simply because they sell. They must also share audiences with aggregators devoted to the business of disseminating content produced by digital news publishers, blogs and RSS feeds, which is chosen on the basis of search engine algorithms, the votes of users or the preferences of readers.
    [Show full text]
  • "Your Car Matters" Tom Dwyer Automotive Services- Spring 2009 Newsletter
    "Your Car Matters" Tom Dwyer Automotive Services- Spring 2009 Newsletter Tom's Tidbits Spring fever may be in the air, but the financial rain continues to fall. This Spring's newsletter continues our attempt to help our clients through this c h al le n ging time. Drew's Kitchen tells you what to do when you just can't eat another bowl of Ramen noodles, our Client Profile tells you about one of our clients who decided that right now is the best time ever to start a new business, and our website has some stories on recession- proofing your job. Kidding aside, we do actually think that the recession will eventually turn around, so looking to the future we examine where our power will come from as world energy demand triples. Finally, w e ' l l tell you about a way to filter the internet down to just the stuff you actually need, and I'll give you a peek at the stuff on the web that I actually need. It may not get you out of the recession, but hopefully it will bring you a smile while you're here. Make a great day, Quarterly Coupons Birds are singing, flowers are blooming, and money's just falling from the sky! WILDCARD- Any Coupon You Want!- Are you regretting not using that third coupon from Summer of 2004? Opportunity knocks again! Go to www.TomDwyer.com, click the "Newsletters" tab, click any newsletter, and use any coupon you find! Coupon must be presented at time of appointment, not combinable with other offers.
    [Show full text]
  • RSS As Medium for Information and Communication Technology
    RSS as medium for information and communication technology Alen Šimec, Mia Čarapina, Sanja Duk Tehnical Polytehnic in Zagreb/Vrbik 8, 10000 Zagreb, Croatia [email protected], [email protected], [email protected] Abstract - In today's modern time we exchange thousands of information with the environment. Over time various forms of information exchange were developed in order to facilitate our daily lives. Today we can use a different approach to exchange information such as correspondence by mail, telegraph, telephone, radio television and internet, etc. In order to preserve data and information they were transferred to a physical media such as paper, and we can conclude that it is a matter of technology that man has evolved over time to facilitate daily life. A term that encompasses the application of such technology is called Information and Communication Technology (English Information and Communication Technology (ICT)). Today, when we use the term information and communication technologies (hereinafter referred to as ICT) primarily we allude to a new technology. With development of ICT came an increasing need for mass information transfer, as for individuals, as for many people. New technologies develop because of the need for a simple way to inform. One such technology that is becoming more popular is the RSS. RSS stands for Really Simple Syndication or Rich Summary Site (at the very beginning of creation, RDF Site Summary). RSS content can be traced through browser or RSS aggregator. RSS aggregator comes in the form of client applications or web application accessed via the web browser. RSS aggregators have the function of collecting titles, descriptions and links to subscribed topics, mostly blogs, podcasts, accounts, news, etc.
    [Show full text]
  • Design and Development of a Learning-Augmented RSS Feed Reader Program “The Mindful Reader”
    Project Number: DCB-0802 Design and Development of a Learning-Augmented RSS Feed Reader Program “The Mindful Reader” A Major Qualifying Project Report: submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE in partial fulfillment of the requirements for the Degree of Bachelor of Science By _________________________________ Chris Drouin Date: April 29, 2009 _________________________________ Professor David C. Brown (CS) Abstract The Mindful Reader Project centers on the design and development of a machine learning- augmented newsfeed aggregation application. It seeks to reduce the time necessary for users to find interesting newsfeed articles, by building a user interest model from implicit and explicit article ratings and applying that model to rank incoming articles based on predicted user interest. The software was developed using code from the RSSOwl project; in tests, the user interest model grew more accurate with time. ii Executive Summary Feed aggregator software and services, such as RSSOwl and Google Reader, can be used to subscribe to websites and obtain streaming updates regarding new articles or items posted to those websites. These existing aggregators follow an email-client-like design that can make it time- consuming and inconvenient to manage high volumes of incoming articles. The Mindful Reader project aims to solve this problem by modeling user interests and using that model to rate and filter incoming articles. As reducing user fatigue is an important goal of this project, the Mindful Reader builds its user interest model in part through observation of normal user behavior, rather than requiring explicit user judgment of every viewed article. It uses metrics established by the earlier Curious Browser projects to measure implicit user interest in article content, while still allowing the user to train the system by providing explicit content ratings.
    [Show full text]
  • Measuring News Consumption in a Digital Era
    1 PEW RESEARCH CENTER FOR RELEASE DECEMBER 8, 2020 Measuring News Consumption in a Digital Era As news outlets morph and multiply, both surveys and passive data collection tools face challenges BY Michael Barthel, Amy Mitchell, Dorene Asare-Marfo, Courtney Kennedy, and Kirsten Worden FOR MEDIA OR OTHER INQUIRIES: Amy Mitchell, Director, Journalism Research Michael Barthel, Senior Researcher Hannah Klein, Communications Manager Rachel Weisel, Senior Communications Manager 202.419.4372 www.pewresearch.org RECOMMENDED CITATION Pew Research Center, December, 2020, “Measuring News Consumption in a Digital Era” www.pewresearch.org 2 PEW RESEARCH CENTER About Pew Research Center Pew Research Center is a nonpartisan fact tank that informs the public about the issues, attitudes and trends shaping the world. It does not take policy positions. The Center conducts public opinion polling, demographic research, content analysis and other data-driven social science research. It studies U.S. politics and policy; journalism and media; internet, science and technology; religion and public life; Hispanic trends; global attitudes and trends; and U.S. social and demographic trends. All of the Center’s reports are available at www.pewresearch.org. Pew Research Center is a subsidiary of The Pew Charitable Trusts, its primary funder. This report was made possible by The Pew Charitable Trusts, which received support from the John S. and James L. Knight Foundation. © Pew Research Center 2020 www.pewresearch.org 3 PEW RESEARCH CENTER How we did this This report is the culmination of a yearlong study into Americans’ news consumption habits in an era of rapidly evolving media technology and a look at how researchers can measure those news habits.
    [Show full text]
  • FEEDLY for Ios
    A7B39TUR Testing mobile application FEEDLY for iOS Cognitive walkthrough and heuristic evaluation Feedly Roman Shkola A7B39TUR Obsah 1. Introduction 3 1.1 About application 3 1.1.1 What is Feedly 3 1.1.2 History 3 1.1.3 What Feedly is used for and why 4 1.1.4 Target group of users 4 2. Use Case 5 2.1 What is the «Use Case» 5 2.2 Feedly’s Use Case 5 2.3 Indexing 6 3. Test methods 7 3.1 Usability inspection 7 3.1.1 Cognitive walkthrough 7 3.1.2 Heuristic evaluation 7 4. App testing 8 4.1 Sign in with existing account 8 4.1.1 Cognitive walkthrough results 9 4.1.2 Heuristic evaluation results 9 4.2 Searching items 10 4.2.1 Cognitive walkthrough results 11 4.2.2 Heuristic evaluation results 11 4.3 Add items to collection 12 4.3.1 Cognitive walkthrough results 12 4.3.2 Heuristic evaluation results 13 4.4 Items sharing 13 4.4.1 Cognitive walkthrough results 15 4.4.2 Heuristic evaluation results 15 5. Summary 16 5.1 Findings 16 5.2 Conclusion 16 A7B39TUR 1. Introduction 1.1 About application 1.1.1 What is Feedly Feedly is a news aggregator application for various web browsers and mobile devices running iOS and Android, also available as a cloud-based service. It compiles news feeds from a variety of online sources for the user to customize and share with others. Feedly was first released by DevHD in 2008.
    [Show full text]