Writing History in the Digital Age
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Analysis of New Dot-Com Bubble in U.S. Stock Market
ANALYSIS OF NEW DOT-COM BUBBLE IN U.S. STOCK MARKET FOR PERIOD 1995 - 2012 By Renate Shafrila Dwima Pranandya ID No. 014200900131 A Thesis presented to the Faculty of Economics President University in partial fulfillment of the requirements for Bachelor Degree in Economics Major in Management March 2013 THESIS ADVISER RECOMMENDATION LETTER This thesis entitled “Analysis of New Dot-com Bubble in U.S. Stock Market for Period 1995 – 2012” prepared and submitted by Renate Shafrila Dwima Pranandya in partial fulfillment of the requirements for the degree of Bachelor of Science in the Faculty of Economics has been reviewed and found to have satisfied the requirements for a thesis fit to be examined. I therefore recommend this thesis for Oral Defense. Cikarang, Indonesia, March 20, 2013 Acknowledged by, Recommended by, Irfan Habsjah, MBA, CMA Ir. Farida Komalasari, M.Si. Head of Management Study Program Student Advisor i DECLARATION OF ORIGINALITY I declare that this thesis, entitled “Analysis of New Dot-com Bubble in U.S. Stock Market for Period 1995 – 2012” is, to the best of my knowledge and belief, an original piece of work that has not been submitted, either in whole or in part, to another university to obtain a degree. Cikarang, Indonesia, March 20, 2013 Renate Shafrila Dwima Pranandya ii PANEL OF EXAMINERS APPROVAL SHEET The Panel of Examiners declare that the thesis entitled “Analysis of New Dot-com Bubble in U.S. Stock Market for Period 1995 – 2012” that was submitted by Renate Shafrila Dwima Pranandya majoring in Management from the Faculty of Economics was assessed and approved to have passed the Oral Examinations on March 4, 2013. -
4€€€€An Encyclopedia with Breaking News
Wikipedia @ 20 • ::Wikipedia @ 20 4 An Encyclopedia with Breaking News Brian Keegan Published on: Oct 15, 2020 Updated on: Nov 16, 2020 License: Creative Commons Attribution 4.0 International License (CC-BY 4.0) Wikipedia @ 20 • ::Wikipedia @ 20 4 An Encyclopedia with Breaking News Wikipedia’s response to the September 11 attacks profoundly shaped its rules and identity and illuminated a new strategy for growing the project by coupling the supply and demand for information about news. Wikipedia’s breaking news collaborations offer lessons for hardening other online platforms against polarization, disinformation, and other sociotechnical sludge. The web was a very different place for news in the United States between 2001 and 2006. The hanging chads from the 2000 presidential election, the spectacular calamity of 9/11, the unrepentant lies around Operation Iraqi Freedom, and the campy reality television featuring Donald Trump were all from this time. The burst of the dot-com bubble and corporate malfeasance of companies like Enron dampened entrepreneurial spirits, news publishers were optimistically sharing their stories online without paywalls, and blogging was heralded as the future of technology-mediated accountability and participatory democracy. “You” was Time Magazine’s Person of the Year in 2006 because “Web 2.0” platforms like YouTube, MySpace, and Second Life had become tools for “bringing together the small contributions of millions of people and making them matter.”1 Wikipedia was a part of this primordial soup, predating news-feed-mediated engagement, recommender-driven polarization, politicized content moderation, and geopolitical disinformation campaigns. From very early in its history, Wikipedia leveraged the supply and demand for information about breaking news and current events into strategies that continue to sustain this radical experiment in online peer production. -
Internet Economy 25 Years After .Com
THE INTERNET ECONOMY 25 YEARS AFTER .COM TRANSFORMING COMMERCE & LIFE March 2010 25Robert D. Atkinson, Stephen J. Ezell, Scott M. Andes, Daniel D. Castro, and Richard Bennett THE INTERNET ECONOMY 25 YEARS AFTER .COM TRANSFORMING COMMERCE & LIFE March 2010 Robert D. Atkinson, Stephen J. Ezell, Scott M. Andes, Daniel D. Castro, and Richard Bennett The Information Technology & Innovation Foundation I Ac KNOW L EDGEMEN T S The authors would like to thank the following individuals for providing input to the report: Monique Martineau, Lisa Mendelow, and Stephen Norton. Any errors or omissions are the authors’ alone. ABOUT THE AUTHORS Dr. Robert D. Atkinson is President of the Information Technology and Innovation Foundation. Stephen J. Ezell is a Senior Analyst at the Information Technology and Innovation Foundation. Scott M. Andes is a Research Analyst at the Information Technology and Innovation Foundation. Daniel D. Castro is a Senior Analyst at the Information Technology and Innovation Foundation. Richard Bennett is a Research Fellow at the Information Technology and Innovation Foundation. ABOUT THE INFORMATION TECHNOLOGY AND INNOVATION FOUNDATION The Information Technology and Innovation Foundation (ITIF) is a Washington, DC-based think tank at the cutting edge of designing innovation policies and exploring how advances in technology will create new economic opportunities to improve the quality of life. Non-profit, and non-partisan, we offer pragmatic ideas that break free of economic philosophies born in eras long before the first punch card computer and well before the rise of modern China and pervasive globalization. ITIF, founded in 2006, is dedicated to conceiving and promoting the new ways of thinking about technology-driven productivity, competitiveness, and globalization that the 21st century demands. -
Wikipedia's Economic Value
WIKIPEDIA’S ECONOMIC VALUE Jonathan Band and Jonathan Gerafi policybandwidth In the copyright policy debate, proponents of strong copyright protection tend to be dismissive of the quality of freely available content. In response to counter- examples such as open access scholarly publications and advertising-supported business models (e.g., newspaper websites and the over-the-air television broadcasts viewed by 50 million Americans), the strong copyright proponents center their attack on amateur content. In this narrative, YouTube is for cat videos and Wikipedia is a wildly unreliable source of information. Recent studies, however, indicate that the volunteer-written and -edited Wikipedia is no less reliable than professionally edited encyclopedias such as the Encyclopedia Britannica.1 Moreover, Wikipedia has far broader coverage. Britannica, which discontinued its print edition in 2012 and now appears only online, contains 120,000 articles, all in English. Wikipedia, by contrast, has 4.3 million articles in English and a total of 22 million articles in 285 languages. Wikipedia attracts more than 470 million unique visitors a month who view over 19 billion pages.2 According to Alexa, it is the sixth most visited website in the world.3 Wikipedia, therefore, is a shining example of valuable content created by non- professionals. Is there a way to measure the economic value of this content? Because Wikipedia is created by volunteers, is administered by a non-profit foundation, and is distributed for free, the normal means of measuring value— such as revenue, market capitalization, and book value—do not directly apply. Nonetheless, there are a variety of methods for estimating its value in terms of its market value, its replacement cost, and the value it creates for its users. -
Twitter, Myspace and Facebook Demystified - by Ted Janusz
Twitter, MySpace and Facebook Demystified - by Ted Janusz Q: I hear people talking about Web sites like Twitter, MySpace and Facebook. What are they? And, even more importantly, should I be using them to promote oral implantology? First, you are not alone. A recent survey showed that 70 percent of American adults did not know enough about Twitter to even have an opinion. Tools like Twitter, Facebook and MySpace are components of something else you may have heard people talking about: Web 2.0 , a popular term for Internet applications in which the users are actively engaged in creating and distributing Web content. Web 1.0 probably consisted of the Web sites you saw back in the late 90s, which were nothing more than fancy electronic brochures. Web 1.5 would have been something like Amazon or eBay, sites on which one could buy, sell and leave reviews. What Web 3.0 will look like is anybody's guess! Let's look specifically at the three applications that you mentioned. Tweet, Tweet Twitter - "Twitter is like text messaging, only you can also do it from the Web," says Dan Tynan, the author of the Tynan on Technology blog. "Instead of sending a message to just one person, you can send it to thousands of people at once. You can choose to follow anyone's update (called "tweets") simply by clicking the Follow button on their profile, or vice-versa. The only rule is that each tweet can be no longer than 140 characters." Former CEO of Twitter Jack Dorsey once accepted an award for Twitter by saying, "We'd like to thank you in 140 characters or less. -
Modeling and Analyzing Latency in the Memcached System
Modeling and Analyzing Latency in the Memcached system Wenxue Cheng1, Fengyuan Ren1, Wanchun Jiang2, Tong Zhang1 1Tsinghua National Laboratory for Information Science and Technology, Beijing, China 1Department of Computer Science and Technology, Tsinghua University, Beijing, China 2School of Information Science and Engineering, Central South University, Changsha, China March 27, 2017 abstract Memcached is a widely used in-memory caching solution in large-scale searching scenarios. The most pivotal performance metric in Memcached is latency, which is affected by various factors including the workload pattern, the service rate, the unbalanced load distribution and the cache miss ratio. To quantitate the impact of each factor on latency, we establish a theoretical model for the Memcached system. Specially, we formulate the unbalanced load distribution among Memcached servers by a set of probabilities, capture the burst and concurrent key arrivals at Memcached servers in form of batching blocks, and add a cache miss processing stage. Based on this model, algebraic derivations are conducted to estimate latency in Memcached. The latency estimation is validated by intensive experiments. Moreover, we obtain a quantitative understanding of how much improvement of latency performance can be achieved by optimizing each factor and provide several useful recommendations to optimal latency in Memcached. Keywords Memcached, Latency, Modeling, Quantitative Analysis 1 Introduction Memcached [1] has been adopted in many large-scale websites, including Facebook, LiveJournal, Wikipedia, Flickr, Twitter and Youtube. In Memcached, a web request will generate hundreds of Memcached keys that will be further processed in the memory of parallel Memcached servers. With this parallel in-memory processing method, Memcached can extensively speed up and scale up searching applications [2]. -
The Culture of Wikipedia
Good Faith Collaboration: The Culture of Wikipedia Good Faith Collaboration The Culture of Wikipedia Joseph Michael Reagle Jr. Foreword by Lawrence Lessig The MIT Press, Cambridge, MA. Web edition, Copyright © 2011 by Joseph Michael Reagle Jr. CC-NC-SA 3.0 Purchase at Amazon.com | Barnes and Noble | IndieBound | MIT Press Wikipedia's style of collaborative production has been lauded, lambasted, and satirized. Despite unease over its implications for the character (and quality) of knowledge, Wikipedia has brought us closer than ever to a realization of the centuries-old Author Bio & Research Blog pursuit of a universal encyclopedia. Good Faith Collaboration: The Culture of Wikipedia is a rich ethnographic portrayal of Wikipedia's historical roots, collaborative culture, and much debated legacy. Foreword Preface to the Web Edition Praise for Good Faith Collaboration Preface Extended Table of Contents "Reagle offers a compelling case that Wikipedia's most fascinating and unprecedented aspect isn't the encyclopedia itself — rather, it's the collaborative culture that underpins it: brawling, self-reflexive, funny, serious, and full-tilt committed to the 1. Nazis and Norms project, even if it means setting aside personal differences. Reagle's position as a scholar and a member of the community 2. The Pursuit of the Universal makes him uniquely situated to describe this culture." —Cory Doctorow , Boing Boing Encyclopedia "Reagle provides ample data regarding the everyday practices and cultural norms of the community which collaborates to 3. Good Faith Collaboration produce Wikipedia. His rich research and nuanced appreciation of the complexities of cultural digital media research are 4. The Puzzle of Openness well presented. -
A Tempest Or, on the flood of Interest In: Sentiment Analysis, Opinion Mining, and the Computational Treatment of Subjective Language
A Tempest Or, on the flood of interest in: sentiment analysis, opinion mining, and the computational treatment of subjective language Lillian Lee Cornell University http://www.cs.cornell.edu/home/llee “Romance should never begin with sentiment. It should begin with science and end with a settlement.” — Oscar Wilde, An Ideal Husband , that has such people in’t According to a Comscore ’07 report and an ’08 Pew survey: 60% of US residents have done online product research. 15% do so on a typical day. 73%-87% of US readers of online reviews of services say the reviews were significant influences. (more on economics later) But 58% of US internet users report that online information was missing, impossible to find, confusing, and/or overwhelming. Creating technologies that find and analyze reviews would answer a tremendous information need. O brave new world People search for and are affected by online opinions. TripAdvisor, Rotten Tomatoes, Yelp, ... Amazon, eBay, YouTube... blogs, Q&A and discussion sites, ... But 58% of US internet users report that online information was missing, impossible to find, confusing, and/or overwhelming. Creating technologies that find and analyze reviews would answer a tremendous information need. O brave new world, that has such people in’t People search for and are affected by online opinions. TripAdvisor, Rotten Tomatoes, Yelp, ... Amazon, eBay, YouTube... blogs, Q&A and discussion sites, ... According to a Comscore ’07 report and an ’08 Pew survey: 60% of US residents have done online product research. 15% do so on a typical day. 73%-87% of US readers of online reviews of services say the reviews were significant influences. -
Seniors' Stories | Volume 4
Seniors’ Stories Seniors’ Stories Volume 4 Volume Volume 4 FRONT COVER: OPEN CATEGORY WINNING ENTRY Angus Lee Forbes Ilona was named after her Great Grandmother, a truly great woman. Born in Hungary, she arrived in Australia during the onset of World War Two. Ninety years and four generations separate the two. Connected by name, both adore spending time with one another. Acknowledgements This collection of 100 stories is the fourth volume of Seniors Stories written by seniors from throughout NSW. The theme of this year’s edition was positive ageing and each story reflects this theme in its own unique and inspiring way. NSW Seniors Card would like to thank the 100 authors whose stories are published in this volume of Seniors Stories, as well as the many other seniors who contributed to the overwhelming number and quality of stories received. Thanks also to contributing editor, Colleen Parker, Fellowship of Australian Writers NSW, and those involved in the design and printing of the book. Seniors’ Stories – Volume 4 1 A message from the Premier Welcome to the fourth edition of Seniors’ Stories. Older people have an extraordinary influence on our The stories in this publication are a great example community; they are the backbone of many of our of the wisdom, talent and ability of our extraordinary voluntary organisations and play a vital role in our seniors in NSW. families and neighbourhoods. The theme of this year’s book is Positive Ageing Seniors’ Stories is just one way of recognising and and the stories showcase the vast range of positive valuing the experiences of older people in NSW experiences and invaluable contributions that older and building connections between young and old. -
PDF of Paper
The Substantial Interdependence of Wikipedia and Google: A Case Study on the Relationship Between Peer Production Communities and Information Technologies Connor McMahon1*, Isaac Johnson2*, and Brent Hecht2 *indicates co-First Authors; 1GroupLens Research, University of Minnesota; 2People, Space, and Algorithms (PSA) Computing Group, Northwestern University [email protected], [email protected], [email protected] Abstract broader information technology ecosystem. This ecosystem While Wikipedia is a subject of great interest in the com- contains potentially critical relationships that could affect puting literature, very little work has considered Wikipedia’s Wikipedia as much as or more than any changes to internal important relationships with other information technologies sociotechnical design. For example, in order for a Wikipedia like search engines. In this paper, we report the results of two page to be edited, it needs to be visited, and search engines deception studies whose goal was to better understand the critical relationship between Wikipedia and Google. These may be a prominent mediator of Wikipedia visitation pat- studies silently removed Wikipedia content from Google terns. If this is the case, search engines may also play a crit- search results and examined the effect of doing so on partici- ical role in helping Wikipedia work towards the goal of pants’ interactions with both websites. Our findings demon- “[providing] every single person on the planet [with] access strate and characterize an extensive interdependence between to the sum of all human knowledge” (Wales 2004). Wikipedia and Google. Google becomes a worse search en- gine for many queries when it cannot surface Wikipedia con- Also remaining largely unexamined are the inverse rela- tent (e.g. -
PDF Preprint
A Deeper Investigation of the Importance of Wikipedia Links to Search Engine Results NICHOLAS VINCENT, Northwestern University, USA BRENT HECHT, Northwestern University, USA A growing body of work has highlighted the important role that Wikipedia’s volunteer-created content plays in helping search engines achieve their core goal of addressing the information needs of hundreds of millions of people. In this paper, we report the results of an investigation into the incidence of Wikipedia links in search engine results pages (SERPs). Our results extend prior work by considering three U.S. search engines, simulating both mobile and desktop devices, and using a spatial analysis approach designed to study modern SERPs that are no longer just “ten blue links”. We find that Wikipedia links are extremely common in important search contexts, appearing in 67-84% of desktop SERPs for common and trending queries, but less often for medical queries. Furthermore, we observe that Wikipedia links often appear in “Knowledge Panel” SERP elements and are in positions visible to users without scrolling, although Wikipedia appears less often and in less prominent positions on mobile devices. Our findings reinforce the complementary notions that (1) Wikipedia content and research has major impact outside of the Wikipedia domain and (2) powerful technologies like search engines are highly reliant on free content 4 created by volunteers. CCS Concepts: • Human-centered computing → Empirical studies in collaborative and social computing KEYWORDS Wikipedia; search engines; user-generated content; data leverage ACM Reference Format: Nicholas Vincent and Brent Hecht. 2021. A Deeper Investigation of the Importance of Wikipedia Links to Search Engine Results. -
Why Are Some Websites Researched More Than Others? a Review of Research Into the Global Top Twenty Mike Thelwall
Why are some websites researched more than others? A review of research into the global top twenty Mike Thelwall How to cite this article: Thelwall, Mike (2020). “Why are some websites researched more than others? A review of research into the global top twenty”. El profesional de la información, v. 29, n. 1, e290101. https://doi.org/10.3145/epi.2020.ene.01 Manuscript received on 28th September 2019 Accepted on 15th October 2019 Mike Thelwall * https://orcid.org/0000-0001-6065-205X University of Wolverhampton, Statistical Cybermetrics Research Group Wulfruna Street Wolverhampton WV1 1LY, UK [email protected] Abstract The web is central to the work and social lives of a substantial fraction of the world’s population, but the role of popular websites may not always be subject to academic scrutiny. This is a concern if social scientists are unable to understand an aspect of users’ daily lives because one or more major websites have been ignored. To test whether popular websites may be ignored in academia, this article assesses the volume and citation impact of research mentioning any of twenty major websites. The results are consistent with the user geographic base affecting research interest and citation impact. In addition, site affordances that are useful for research also influence academic interest. Because of the latter factor, however, it is not possible to estimate the extent of academic knowledge about a site from the number of publications that mention it. Nevertheless, the virtual absence of international research about some globally important Chinese and Russian websites is a serious limitation for those seeking to understand reasons for their web success, the markets they serve or the users that spend time on them.