The Impact of Javascript on Archivability

Total Page:16

File Type:pdf, Size:1020Kb

The Impact of Javascript on Archivability Int J Digit Libr DOI 10.1007/s00799-015-0140-8 The impact of JavaScript on archivability Justin F. Brunelle · Mat Kelly · Michele C. Weigle · Michael L. Nelson Received: 7 November 2013 / Revised: 12 January 2015 / Accepted: 14 January 2015 © Springer-Verlag Berlin Heidelberg 2015 Abstract As web technologies evolve, web archivists work 12.0 % from 2005 to 2012. We also show that JavaScript to adapt so that digital history is preserved. Recent advances is responsible for 33.2 % more missing resources in 2012 in web technologies have introduced client-side executed than in 2005. This shows that JavaScript is responsible for scripts (Ajax) that, for example, load data without a change an increasing proportion of the embedded resources unsuc- in top level Universal Resource Identifier (URI) or require cessfully loaded by mementos. JavaScript is also responsible user interaction (e.g., content loading via Ajax when the for 52.7 % of all missing embedded resources in our study. page has scrolled). These advances have made automat- ing methods for capturing web pages more difficult. In an Keywords Web architecture · Web archiving · effort to understand why mementos (archived versions of Digital preservation live resources) in today’s archives vary in completeness and sometimes pull content from the live web, we present a study of web resources and archival tools. We used a collection of 1 Introduction URIs shared over Twitter and a collection of URIs curated by Archive-It in our investigation. We created local archived ver- How well can we archive the web? This is a question that sions of the URIs from the Twitter and Archive-It sets using is becoming increasingly important and more difficult to WebCite, wget, and the Heritrix crawler. We found that only answer. Additionally, this question has significant impact on 4.2 % of the Twitter collection is perfectly archived by all of web users [40,43] and commercial and government compli- these tools, while 34.2 % of the Archive-It collection is per- ance [52,53,66]. fectly archived. After studying the quality of these mementos, The web has gone through a gradient of changes fueled by we identified the practice of loading resources via JavaScript increasing user demand for interactivity. Early websites were (Ajax) as the source of archival difficulty. Further, we show relatively static, while continued adoption of web technolo- that resources are increasing their use of JavaScript to load gies has made the pages personalized and more interactive. embedded resources. By 2012, over half (54.5 %) of pages JavaScript, which executes on the client, provides additional use JavaScript to load embedded resources. The number of features for the web user, enabling or increasing interactivity, embedded resources loaded via JavaScript has increased by client-side state changes, and personalized representations. These additional features offer an enhanced browsing expe- J. F. Brunelle (B) · M. Kelly · M. C. Weigle · M. L. Nelson rience for the user. Department of Computer Science, Old Dominion University, Norfolk, VA 23529, USA JavaScript has enabled a wide-scale migration from web e-mail: [email protected] pages to web applications. This migration continued with the M. Kelly introduction of Ajax (first introduced in 2005 [28]), which e-mail: [email protected] combined multiple technologies to give web pages the abil- M. C. Weigle ity to perform asynchronous client–server interactions after e-mail: [email protected] the HTML is loaded. The first wide-scale implementation of M. L. Nelson Ajax was in Google Maps in 2005, but Ajax was officially e-mail: [email protected] added as a standard in 2006 [70]. While archival tools per- 123 J. F. Brunelle et al. (a) (b) (c) Fig. 1 Google Maps as it exists live and as a memento a live version, b archived Dec. 1, 2012, c archived Apr. 30, 2012 formed satisfactorily when the web consisted almost entirely We discuss how these technologies impact archivability in of static resources, archival tools (e.g., Heritrix [51,65], Sect. 9. Many culturally significant artifacts (such as Google WebCite [22], and the wget application [29]) cannot capture Maps, shown in Fig. 1) are unable to be unarchived. Even the more complex, feature-rich pages making up today’s web. if resources are archived, they sometimes become tempo- The ease of archiving a web page (henceforth the archiv- rally inconsistent because JavaScript may load content from ability) is impacted by the migration from web pages to web the live web (referred to as the live web “leaking” into the applications. Over the years the web has shifted from a web archive and discussed in more detail in Sect. 6.4) instead of of documents to a web of applications, and these applica- only loading temporally appropriate content. tions require execution by the client to render what the user Throughout this paper we use Memento Framework ter- experiences. Client-side rendering has not historically been minology. Memento [68] is a framework that standardizes part of the archival process. Even if it were, the applica- Web archive access and terminology. Original (or live web) tions depend on client-side inputs that increase the variability resources are identified by URI-R, and archived versions of between what is archived and what the user experienced [60]. URI-Rs are called mementos and are identified by URI-M. Because of the advancement and increasing adoption of Memento TimeMaps are machine-readable lists of memen- web technologies, the web is composed of an increasing pro- tos (at the level of single-archives or aggregation-of-archives) portion of personalized resources that are difficult to archive. sorted by archival date. 123 The impact of Javascript on archivability This paper makes two contributions. The first contribution refers to the final representation that is not fully realized is a study of live resources and mementos to provide insight and constructed until after the client-side representation is into what makes a resource archivable. We studied two col- rendered; the representation’s final state is deferred until after lections of resources and measured them for archivability. client-side code and events have finished executing. Heritrix, WebCite, and wget are all used to capture a set of As we have mentioned, web browsers use an engine to resources shared via Twitter and a set of Archive-It curated run client-side JavaScript. Web crawlers often do not pos- resources with the intent to determine how well each tool sess such an engine and, therefore, do not have the ability archives the resources. These tools, collections, and result- to execute JavaScript during the archival process. In other ing mementos are compared, and the results are discussed in words, archival crawlers are not equipped with the features depth to show that Twitter resources are less archivable than necessary to execute JavaScript. If Ajax requests additional Archive-It resources, and Heritrix is the best archival tool resources to be embedded in the representation, crawlers will of the three observed. In culmination, the first contribution be unaware of the extra resources required to properly ren- provides an analysis of how well we can archive the web. der the representation and, ultimately, will not archive these The second contribution is a study of how archivability has embedded resources. changed over time. We study the mementos of the URI-Rs Because most web crawlers do not have the ability to exe- in our collections for archival quality and cause of missing cute embedded JavaScript or other client-side technologies, embedded resources. We show that resources are becoming the resulting mementos may only be partially operational or less archivable over time due to the increasing prevalence of incomplete. We discuss crawlers’ current capabilities further JavaScript. In short, the second contribution provides insight in Sect. 4. into how well archival tools have performed in the past. Google Maps is an example of a resource with a deferred representation and where the lack of completeness of memen- tos is easy to observe. Figure 1a shows the live page, con- 2 Deferred representations taining multiple interactive UI elements. The map in the mid- dle is draggable, allowing the user to pan. An archived ver- According to the web architecture [10,23,32], servers deliver sion (from April 30, 2012) of this page in Fig. 1c is missing web resources to requesting clients such as a web browser UI elements and functionality (circled), and the interaction (controlled by a user, such as Firefox [24]) or a web crawler (e.g., panning and zooming) does not function. This is due (e.g., Heritrix or other robot). The content delivered to the to resources that would be loaded when the user clicks, but client by the server is the representation of the web resource that are not preserved by the crawler. Figure 1b shows an (e.g., HTML, PDF). The representation may also include a archived (December 1, 2012) version that gives the façade client-side programming language such as JavaScript which of functionality when, in fact, resources on the live web are will execute when the representation is rendered by a web being loaded. browser and the client-side code is run by a local compiler Figure 2 outlines the role Ajax plays in loading a resource or—in the case of JavaScript [25]—an engine. like Google Maps. When a crawler fetches a web page (1), Executing a technology like JavaScript on the client can it waits for the page and all of the embedded resources to potentially cause the representation to change with or without load to consider it fully rendered. At this point, the crawler subsequent requests to a server for additional resources.
Recommended publications
  • Modeling Popularity and Reliability of Sources in Multilingual Wikipedia
    information Article Modeling Popularity and Reliability of Sources in Multilingual Wikipedia Włodzimierz Lewoniewski * , Krzysztof W˛ecel and Witold Abramowicz Department of Information Systems, Pozna´nUniversity of Economics and Business, 61-875 Pozna´n,Poland; [email protected] (K.W.); [email protected] (W.A.) * Correspondence: [email protected] Received: 31 March 2020; Accepted: 7 May 2020; Published: 13 May 2020 Abstract: One of the most important factors impacting quality of content in Wikipedia is presence of reliable sources. By following references, readers can verify facts or find more details about described topic. A Wikipedia article can be edited independently in any of over 300 languages, even by anonymous users, therefore information about the same topic may be inconsistent. This also applies to use of references in different language versions of a particular article, so the same statement can have different sources. In this paper we analyzed over 40 million articles from the 55 most developed language versions of Wikipedia to extract information about over 200 million references and find the most popular and reliable sources. We presented 10 models for the assessment of the popularity and reliability of the sources based on analysis of meta information about the references in Wikipedia articles, page views and authors of the articles. Using DBpedia and Wikidata we automatically identified the alignment of the sources to a specific domain. Additionally, we analyzed the changes of popularity and reliability in time and identified growth leaders in each of the considered months. The results can be used for quality improvements of the content in different languages versions of Wikipedia.
    [Show full text]
  • Cultural Anthropology and the Infrastructure of Publishing
    CULTURAL ANTHROPOLOGY AND THE INFRASTRUCTURE OF PUBLISHING TIMOTHY W. ELFENBEIN Society for Cultural Anthropology This essay is based on an interview conducted by Matt Thompson, portions of which originally appeared on the blog Savage Minds the week after Cultural Anthropology relaunched as an open-access journal.1 Herein, I elaborate on my original comments to provide a nuts-and-bolts account of what has gone into the journal’s transition. I am particularly concerned to lay bare the publishing side of this endeavor—of what it means for a scholarly society to take on publishing responsibilities—since the challenges of pub- lishing, and the labor and infrastructure involved, are invisible to most authors and readers. Matt Thompson: When did the Society for Cultural Anthropology (SCA) de- cide to go the open-access route and what was motivating them? Tim Elfenbein: Cultural Anthropology was one of the first American Anthropo- logical Association (AAA) journals to start its own website. Kim and Mike Fortun were responsible for the initial site. They wanted to know what extra materials they could put on a website that would supplement the journal’s articles. That experience helped germinate the idea that there was more that the journal could do itself. The Fortuns are also heavily involved in science and technology studies 2 (STS), where discussions about open access have been occurring for a long time. When Anne Allison and Charlie Piot took over as editors of the journal in 2011, they continued to push for an open-access alternative in our publishing program. Although the SCA and others had been making the argument for transitioning at CULTURAL ANTHROPOLOGY, Vol.
    [Show full text]
  • Wikipedia and Medicine: Quantifying Readership, Editors, and the Significance of Natural Language
    JOURNAL OF MEDICAL INTERNET RESEARCH Heilman & West Original Paper Wikipedia and Medicine: Quantifying Readership, Editors, and the Significance of Natural Language James M Heilman1, BSc, MD, CCFP(EM); Andrew G West2, PhD, MS Eng 1Faculty of Medicine, Department of Emergency Medicine, University of British Columbia, Vancouver, BC, Canada 2Verisign Labs (Verisign, Inc.), Reston, VA, United States Corresponding Author: James M Heilman, BSc, MD, CCFP(EM) Faculty of Medicine Department of Emergency Medicine University of British Columbia 2194 Health Sciences Mall, Unit 317 Vancouver, BC, V6T1Z3 Canada Phone: 1 4158306381 Fax: 1 6048226061 Email: [email protected] Abstract Background: Wikipedia is a collaboratively edited encyclopedia. One of the most popular websites on the Internet, it is known to be a frequently used source of health care information by both professionals and the lay public. Objective: This paper quantifies the production and consumption of Wikipedia's medical content along 4 dimensions. First, we measured the amount of medical content in both articles and bytes and, second, the citations that supported that content. Third, we analyzed the medical readership against that of other health care websites between Wikipedia's natural language editions and its relationship with disease prevalence. Fourth, we surveyed the quantity/characteristics of Wikipedia's medical contributors, including year-over-year participation trends and editor demographics. Methods: Using a well-defined categorization infrastructure, we identified medically pertinent English-language Wikipedia articles and links to their foreign language equivalents. With these, Wikipedia can be queried to produce metadata and full texts for entire article histories. Wikipedia also makes available hourly reports that aggregate reader traffic at per-article granularity.
    [Show full text]
  • Web Archiving and You Web Archiving and Us
    Web Archiving and You Web Archiving and Us Amy Wickner University of Maryland Libraries Code4Lib 2018 Slides & Resources: https://osf.io/ex6ny/ Hello, thank you for this opportunity to talk about web archives and archiving. This talk is about what stakes the code4lib community might have in documenting particular experiences of the live web. In addition to slides, I’m leading with a list of material, tools, and trainings I read and relied on in putting this talk together. Despite the limited scope of the talk, I hope you’ll each find something of personal relevance to pursue further. “ the process of collecting portions of the World Wide Web, preserving the collections in an archival format, and then serving the archives for access and use International Internet Preservation Coalition To begin, here’s how the International Internet Preservation Consortium or IIPC defines web archiving. Let’s break this down a little. “Collecting portions” means not collecting everything: there’s generally a process of selection. “Archival format” implies that long-term preservation and stewardship are the goals of collecting material from the web. And “serving the archives for access and use” implies a stewarding entity conceptually separate from the bodies of creators and users of archives. It also implies that there is no web archiving without access and use. As we go along, we’ll see examples that both reinforce and trouble these assumptions. A point of clarity about wording: when I say for example “critique,” “question,” or “trouble” as a verb, I mean inquiry rather than judgement or condemnation. we are collectors So, preambles mostly over.
    [Show full text]
  • Web Archiving for Academic Institutions
    University of San Diego Digital USD Digital Initiatives Symposium Apr 23rd, 1:00 PM - 4:00 PM Web Archiving for Academic Institutions Lori Donovan Internet Archive Mary Haberle Internet Archive Follow this and additional works at: https://digital.sandiego.edu/symposium Donovan, Lori and Haberle, Mary, "Web Archiving for Academic Institutions" (2018). Digital Initiatives Symposium. 4. https://digital.sandiego.edu/symposium/2018/2018/4 This Workshop is brought to you for free and open access by Digital USD. It has been accepted for inclusion in Digital Initiatives Symposium by an authorized administrator of Digital USD. For more information, please contact [email protected]. Web Archiving for Academic Institutions Presenter 1 Title Senior Program Manager, Archive-It Presenter 2 Title Web Archivist Session Type Workshop Abstract With the advent of the internet, content that institutional archivists once preserved in physical formats is now web-based, and new avenues for information sharing, interaction and record-keeping are fundamentally changing how the history of the 21st century will be studied. Due to the transient nature of web content, much of this information is at risk. This half-day workshop will cover the basics of web archiving, help attendees identify content of interest to them and their communities, and give them an opportunity to interact with tools that assist with the capture and preservation of web content. Attendees will gain hands-on web archiving skills, insights into selection and collecting policies for web archives and how to apply what they've learned in the workshop to their own organizations. Location KIPJ Room B Comments Lori Donovan works with partners and the Internet Archive’s web archivists and engineering team to develop the Archive-It service so that it meets the needs of memory institutions.
    [Show full text]
  • Web Archiving in the UK: Current
    CITY UNIVERSITY LONDON Web Archiving in the UK: Current Developments and Reflections for the Future Aquiles Ratti Alencar Brayner January 2011 Submitted in partial fulfillment of the requirements for the degree of MSc in Library Science Supervisor: Lyn Robinson 1 Abstract This work presents a brief overview on the history of Web archiving projects in some English speaking countries, paying particular attention to the development and main problems faced by the UK Web Archive Consortium (UKWAC) and UK Web Archive partnership in Britain. It highlights, particularly, the changeable nature of Web pages through constant content removal and/or alteration and the evolving technological innovations brought recently by Web 2.0 applications, discussing how these factors have an impact on Web archiving projects. It also examines different collecting approaches, harvesting software limitations and how the current copyright and deposit regulations in the UK covering digital contents are failing to support Web archive projects in the country. From the perspective of users’ access, this dissertation offers an analysis of UK Web archive interfaces identifying their main drawbacks and suggesting how these could be further improved in order to better respond to users’ information needs and access to archived Web content. 2 Table of Contents Abstract 2 Acknowledgements 5 Introduction 6 Part I: Current situation of Web archives 9 1. Development in Web archiving 10 2. Web archiving: approaches and models 14 3. Harvesting and preservation of Web content 19 3.1 The UK Web space 19 3.2 The changeable nature of Web pages 20 3.3 The evolution of Web 2.0 applications 23 4.
    [Show full text]
  • Archiving Web Content Jean-Christophe Peyssard
    Archiving Web Content Jean-Christophe Peyssard To cite this version: Jean-Christophe Peyssard. Archiving Web Content. École thématique. Archiving Web Content, American University of Beirut, Beirut, Lebanon, Lebanon. 2019. cel-02130558 HAL Id: cel-02130558 https://halshs.archives-ouvertes.fr/cel-02130558 Submitted on 15 May 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Distributed under a Creative Commons Attribution| 4.0 International License American University of Beirut May 3rd 2019 ARCHIVING WEB CONTENT 2019 Jean-Christophe Peyssard Head of Digital Humanities Institut français du Proche-Orient (Ifpo) [email protected] Who Am I ? Research Ingineer from CNRS at the French Institute of the Near East, Head of Digital Humanities . Ifpo : http://www.ifporient.org/jean-christophe-peyssard/ . Linkedin : http://fr.linkedin.com/pub/jean-christophe-peyssard/22/705/782/ . ORCID : 0000-0002-8503-2217 . Google Scholar : https://scholar.google.com/citations?user=32cZHPsAAAAJ . HAL : https://cv.archives-ouvertes.fr/jcpeyssard Institut français
    [Show full text]
  • Web Archiving & Digital Libraries
    Proceedings o f th e 201 8 Web Archiving & Digital Libraries Workshop June 6, 2018 Fort Worth, Texas Martin Klein Edward A. Fox Zhiwu Xie Edito rs Web Archiving And Digital Libraries 2018 W ADL 2018 homepage Web Archiving and Digital Libraries Workshop at JCDL 2018 (http://2018.jcdl.org) Fort Worth, TX, USA Please see the approved WADL 2018 workshop proposal. Please also see last year's WADL 2017 homepage and the homepage of the 2016 edition of WADL. That workshop led in part to a special issue of International Journal on Digital Libraries. We fully intend to publish a very similar IJDL issue based on WADL 2018 contributions. SCHEDULE: Featured Talk by Weigle, Michele C. ([email protected]): Enabling Personal Use of Web Archives W ednesday, June 6, 10:30am-5pm Time Activity, Presenters/Authors Title of Presentation 10:30 Organizers and everyone speaking Opening, Introductions 10:45 Zhiwu Xie et al. IMLS project panel John Berlin, Michael Nelson and Michele Swimming In A Sea Of JavaScript, Or: How I Learned 11:45 Weigle To Stop Worrying And Love High-Fidelity Replay 12:15 Get boxes and return Lunch A Study of Historical Short URLs in Event Collections 1:00 Liuqing Li and Edward Fox of Tweets 1:30 Keynote by Michele Weigle Enabling Personal Use of Web Archives Libby Hemphill, Susan Leonard and Margaret 2:30 Developing a Social Media Archive at ICPSR Hedstrom Posters: Littman Justin; Sawood Alam, Mat Supporting social media research at scale; A Survey of 3:00 Kelly, Michele Weigle and Michael Nelson Archival Replay Banners 3:15 Discussions around posters Break Mohamed Aturban, Michael Nelson and 3:30 It is Hard to Compute Fixity on Archived Web Pages Michele Weigle Mat Kelly, Sawood Alam, Michael Nelson and Client-Assisted Memento Aggregation Using the Prefer 4:00 Michele Weigle Header 4:30 Closing discussion Plans for future activities and collaborations Description: The 2018 edition of the Workshop on Web Archiving and Digital Libraries (WADL) will explore the integration of Web archiving and digital libraries.
    [Show full text]
  • Creating an Online Archive for LGBT History by Ross Burgess
    Creating an online archive for LGBT history By Ross Burgess Presented at the third academic conference, ‘What is & How to Do LGBT History: Approaches, Methods & Subjects’ Manchester, February 2016 In July 2011 I heard about a new Wikipedia-style project to record Britain’s LGBT history, and immediately knew it was something I wanted to help with. I had been editing Wikipedia articles for about seven years, so I was very familiar with the wiki approach to building an online database; and I also had an interest in LGBT history: for instance in 2004 my partner and I had staged re-enactments of the mediaeval church’s same-sex union ceremonies – adelphopoiesis – as part of the campaign to support the introduction of civil partnerships.1 So when Jonathan Harbourne asked for volunteer editors for his new LGBT History wiki, I signed up at once. The project’s title includes ‘LGBT’ (lesbian, gay, bisexual, transgender) which is probably the most convenient and widely understood abbreviation for our community. But its scope could be better defined as ‘LGBT+’ – it definitely includes intersex people, those who classify themselves as queer, non-binary, pansexual, etc, etc. Jonathan’s project was a virtual ‘time capsule’, a resource for people and aca- demics in years to come. The image of the Figure 1. The original logo time capsule has been part of the project’s logo from the outset. He said at the time, In another ten or twenty years, I don’t want people to forget the struggle and the fight that won us our equalities and freedoms.
    [Show full text]
  • Web-Archiving 01000100 01010000 Maureen Pennock 01000011 01000100 DPC Technology Watch Report 13-01 March 2013 01010000 01000011 01000100 01010000
    01000100 01010000 01000011 Web-Archiving 01000100 01010000 Maureen Pennock 01000011 01000100 DPC Technology Watch Report 13-01 March 2013 01010000 01000011 01000100 01010000 Series editors on behalf of the DPC 01000011 Charles Beagrie Ltd. Principal Investigator for the Series 01000100 Neil Beagrie 01010000 01000011DPC Technology Watch Series © Digital Preservation Coalition 2013 and Maureen Pennock 2013 Published in association with Charles Beagrie Ltd. ISSN: 2048 7916 DOI: http://dx.doi.org/10.7207/twr13-01 All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, without the prior permission in writing from the publisher. The moral right of the author has been asserted. First published in Great Britain in 2013 by the Digital Preservation Coalition Foreword The Digital Preservation Coalition (DPC) is an advocate and catalyst for digital preservation, ensuring our members can deliver resilient long-term access to digital content and services. It is a not-for-profit membership organization whose primary objective is to raise awareness of the importance of the preservation of digital material and the attendant strategic, cultural and technological issues. It supports its members through knowledge exchange, capacity building, assurance, advocacy and partnership. The DPC’s vision is to make our digital memory accessible tomorrow. The DPC Technology Watch Reports identify, delineate, monitor and address topics that have a major bearing on ensuring our collected digital memory will be available tomorrow. They provide an advanced introduction in order to support those charged with ensuring a robust digital memory, and they are of general interest to a wide and international audience with interests in computing, information management, collections management and technology.
    [Show full text]
  • 0A Front Matter-Copyright-Approval-Title Page
    Copyright by James Joseph Brown, Jr. 2009 The Dissertation Committee for James Joseph Brown, Jr. certifies that this is the approved version of the following dissertation: Hospitable Texts Committee: _________________________________ D. Diane Davis, Supervisor _________________________________ N. Katherine Hayles _________________________________ Cynthia Haynes _________________________________ Clay Spinuzzi _________________________________ Margaret Syverson Hospitable Texts By James Joseph Brown, Jr., B.S., M.A. Dissertation Presented to the Faculty of the Graduate School of The University of Texas at Austin in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy The University of Texas at Austin May 2009 Acknowledgements The trail of citations that led to this project is long and filled with the most gracious and brilliant scholars I could ever have hoped to work with. First and foremost, Diane Davis (DDD) was willing to read countless drafts of these chapters. As DDD will tell you, I work in threes. Three drafts of each chapter, numerous conversations, and various DDD urgings to “slow down” brought you the project that follows. DDD taught me how to read, and I can only hope that this dissertation is a reflection of her ability to perform and teach a gracious and careful mode of scholarship. My entire dissertation committee was immensely helpful throughout the process. This project would never have gotten off the ground without Clay Spinuzzi’s help during the prospectus stages. Feedback at various stages from Kate Hayles, Cynthia Haynes, and Peg Syverson was invaluable. In addition, all of these scholars can find their own stamp on this project at various moments. Their theories and wisdom are the stuff out of which my writing has emerged.
    [Show full text]
  • List of Epidemics
    List of epidemics This is a list of the biggest known epidemics (including pandemics) caused by an infectious disease. Widespread non-communicable diseases such as cardiovascular disease and cancer are not included. An epidemic is the rapid spread of disease to a large number of people in a given population within a short period of time. For example, in meningococcal infections, an attack rate in excess of 15 cases per 100,000 people for two consecutive weeks is considered an epidemic.[1] 17th-century German "plague panel" depicting the triumph of death. Panels of this kind were placed on the walls of houses to warn against the plague. A plague epidemic raged in Augsburg, Bavaria between 1632 and 1635 Emergency hospital during Spanish Influenza epidemic, Camp Funston, Kansas (1918-20) Death toll Event Date Location Disease Ref. (estimate) Babylon, or Babirus of the Persians, Sanskrit scholars found Influenza 1200 BC Central Asia, records of a disease Unknown [2] epidemic Mesopotamia resembling the Flu and Southern Asia Unknown, possibly Plague of 429–426 Greece, Libya, typhus, typhoid fever or 75,000–100,000 [3][4][5][6] Athens BC Egypt, Ethiopia viral hemorrhagic fever Greece 412 BC (Northern Unknown, possibly 412 BC Unknown [7] epidemic Greece, Roman influenza Republic) 165–180 Unknown, possibly Antonine Plague (possibly Roman Empire 5–10 million [8] smallpox up to 190) Plague of Unknown, possibly 250–266 Europe 1 million+ [9][10] Cyprian smallpox 25–100 million (40– Plague of Europe and Bubonic plague (beginning 541–542 50% of population [11][12][13] Justinian West Asia of First plague pandemic) of Europe) Rome, Roman Plague Bubonic plague (part of 590 Byzantine Unknown [14] of 590 First plague pandemic) Empire Plague of Bubonic plague (part of 627–628 Mesopotamia Unknown [15] Sheroe First plague pandemic) Bubonic plague (part of Plague of 664 664–689 British Isles Unknown [16] First plague pandemic) Byzantine Plague of 698– Empire, West Bubonic plague (part of 698–701 Unknown [17] 701 Asia, Syria, First plague pandemic) Mesopotamia 735–737 2 million (approx.
    [Show full text]