Societal Costs of “Fake News” in the Digital Single Market

Total Page:16

File Type:pdf, Size:1020Kb

Societal Costs of “Fake News” in the Digital Single Market STUDY R equested by the IMCO Committee Societal costs of “fake news” in the Digital Single Market Policy Department for Economic, Scientific and Quality of Life Policies Directorate-General for Internal Policies Author: Prof. Divina Frau-Meigs EN PE 626.087- December 2018 DO NOT DELETE PAGE BREAK Societal costs of “fake news” in the Digital Single Market Abstract This study explores the mechanisms of “fake news” and their societal costs in the Digital Single Market. It describes the risks to the integrity of information and to the integrity of elections. It highlights the roles of the various actors involved in the production and amplification of such information disorders. Finally, it outlines responses that are being tested in different parts of Europe to deal with the issue. The document has been provided by Policy Department A at the request of the European Parliament Committee on the Internal Market and Consumer Protection. This document was commissioned by the European Parliament's Committee on the Internal Market and Consumer Protection. AUTHOR Prof. Dr Divina FRAU-MEIGS, University Sorbonne Nouvelle, Paris, France ADMINISTRATOR RESPONSIBLE Mariusz MACIEJEWSKI EDITORIAL ASSISTANT Irene VERNACOTOLA LINGUISTIC VERSIONS Original: EN ABOUT THE EDITOR Policy departments provide in-house and external expertise to support EP committees and other parliamentary bodies in shaping legislation and exercising democratic scrutiny over EU internal policies. To contact the Policy Department or to subscribe for updates, please write to: Policy Department for Economic, Scientific and Quality of Life Policies European Parliament B-1047 Brussels Email: [email protected] Manuscript completed: November 2018 Date of publication: January 2019 © European Union, 2019 This document is available on the internet at: http://www.europarl.europa.eu/supporting-analyses DISCLAIMER AND COPYRIGHT The opinions expressed in this document are the sole responsibility of the authors and do not necessarily represent the official position of the European Parliament. Reproduction and translation for non-commercial purposes are authorised, provided the source is acknowledged and the European Parliament is given prior notice and sent a copy. For citation purposes, the study should be referenced as: Frau-Meigs, D., Societal costs of “fake news” in the Digital Single Market, Study for the Committee on the Internal Market and Consumer Protection, Policy Department for Economic, Scientific and Quality of Life Policies, European Parliament, Luxembourg, 2018.© Cover image used under licence from Shutterstock.com DELETE IF NOT SHUTTERSTOCK Societal costs of “fake news” in the Digital Single Market CONTENTS LIST OF ABBREVIATIONS 4 LIST OF BOXES 5 LIST OF FIGURES 5 LIST OF TABLES 5 EXECUTIVE SUMMARY 6 1. INTRODUCTION 9 2. SPECIFICITY OF DIGITAL DISINFORMATION AND INTRINSIC MECHANISMS 10 2.1 The disinformation trap-situation 10 2.2. From “fakenews” to malinformation 13 2.3. Major mechanisms 14 3. SOCIETAL COSTS 16 3.1 Fakedom in all sectors 16 3.2. Cost to journalism and risk of harm to integrity of information 18 3.2.1. Concentration of ownership and advertising 18 3.2.2. Convergence between mass media and social media: friends and foes 20 3.3. Cost to integrity of elections and hybrid threats 21 3.3.1. Public relations vs. journalism 21 3.3.2. Vulnerability of elections 23 4. RESPONSES : FROM SOFT TO HARD TO SMART 25 4.1. Self-regulation: fact-checking on the rise 26 4.2. Regulation: towards a transparency agenda 28 4.3. Media and information literacy: smart sense-making 30 5. CONCLUSION: TOWARDS A NEW PHASE OF CO-REGULATION AND RE-INTERMEDIATION 32 REFERENCES 34 REPORTS 35 DOCUMENTARIES 35 PE 626.087 3 IPOL | Policy Department for Economic, Scientific and Quality of Life Policies LIST OF ABBREVIATIONS AMP Accelerated Mobile Pages API Application Programming Interface CBS Corporate Broadcasting Service CPM Cost per Mile CSA Conseil Supérieur de l’Audiovisuel DDoS Distributed Denial Of Service GAFAM Google, Apple, Facebook, Amazon, Microsoft GDPR General Data Protection Regulation H2020 Horizon 2020 work programme IFCN International Fact-Checking Network MAS Media Accountability System MCN Multi Channel Network MIL Media and Information Literacy NATO North Atlantic Treaty Organisation OSCE Organization for Security and Cooperation in Europe PAC Political Action Committee PISA Programme for International Student Assessment PR Public Relations 4 PE 626.087 Societal costs of “fake news” in the Digital Single Market LIST OF BOXES Box 1: The “Donald Cook” Story 9 Box 2: Examples of international initiative in fact-checking 26 Box 3: Examples of MIL good pratctices to counter “fakenews” 31 LIST OF FIGURES Figure 1: The (dis)information ecosystem 12 Figure 2: The three layers of malinformation 12 Figure 3: Two-Stepped Convergence of mainstream mass media and social media platforms 20 LIST OF TABLES Table 1: Social media revenue 11 Table 2: Most shared climate article on social media 18 Table 3: The biggest media corporations in the world 19 Table 4: Spending on political advertising in USA (2008-20) 22 Table 5: Trust in journalism (2018) 27 PE 626.087 5 IPOL | Policy Department for Economic, Scientific and Quality of Life Policies EXECUTIVE SUMMARY The “fake news” phenomenon has come to the centre of public and political attention since 2016, with the suspicion that several electoral events (elections in the U.S. and in France as well as referenda in the UK and Catalonia) have been tampered with. In all cases, the role of social media platforms as major vectors of such disruptive phenomena has been questioned, especially with scandals such as Cambridge Analytica’s use of data from Facebook, and court cases such as Internet Research Agency vs. United States of America. As new elections are approaching in European countries, in particular the European Parliament´s elections in May 2019, it is urgent and necessary to evaluate the extent of the situation and adjust the necessary responses, especially in relation to trust in the Digital Single Market. The purpose of this in-depth analysis, provided at the request of the European Parliament’s Committee on the Internal Market and Consumer Protection, is to set the general framework for understanding the issue of “fake news” and how it can affect trust in the Digital Single Market. The paper offers an explanation of the phenomenon of “fake news” and indicates its evolution, from past processes to new specific incidents, differentiated from neighbouring concepts, such as propaganda or conspiracy. It provides for some elements of categorisation of “fake news” and illustrates their specificity in the XXIst century both with regard to the societal weaknesses that make Europe especially vulnerable to “fake news” and communication innovations that drive them. It redefines “fake news” as “malinformation”, i.e. the triple conjunction of human malevolence, marketing malpractice and engineered malware. It situates malinformation in the family of information disorders such as misinformation, online radicalization and hate speech. The analysis also reflects on the societal costs of this malevolent combination. Among these, there are economic costs but also, most important here, two major societal costs: the integrity of information and the integrity of elections. As regards integrity of information, the paper points to the numerous asymmetries that exist not only between mainstream mass media and social media platforms but also between social media in the U.S. and in Europe. The European digital market has not been able to create search engines or social networks large enough to compete with the American ones. Many asymmetries between American and European companies are immediately apparent: tax obligations, social responsibilities, advertising constraints and public service obligations (like diversity, pluralism, and protection of minors). So nothing hinders the American-based social media from spreading malinformation to increase traffic and thus profits. These asymmetries increase risks of harm to legitimate information as it becomes extremely costly to produce and puts quality reference media under pressure to yield more verification of news, which in turn increases costs and adds time delays. All this takes place in a context of increasingly concentrated ownership and pressure for profits from share-holders and directors who consider information to be an appendage to their larger sectors of activity which can include construction, defence or other economic sectors without long-standing interest or competence in news. As regards the integrity of elections, the paper describes the emergence of “hybrid threats” and the increased capacity of malign foreign intervention into EU affairs as the tension grows between democratic regimes and authoritarian regimes. Various phases of malinformation are described as well as the e-repertoire of strategies that recombine different mechanisms of malinformation into hybrid threats. Among the most characteristic elements are the drip dissemination of topics likely to create viral trends and polarization of topics well before the election time; the hack and delivery of strategic data to cause surprise or panic at determined stages of a campaign; the whitewashing of “fake news” via social media to promote amplification via mass media and conversely, and opportunistically, secret 6 PE 626.087 Societal costs of “fake news” in the Digital Single Market collusion or objective alignment of interests
Recommended publications
  • Cancel Culture: Posthuman Hauntologies in Digital Rhetoric and the Latent Values of Virtual Community Networks
    CANCEL CULTURE: POSTHUMAN HAUNTOLOGIES IN DIGITAL RHETORIC AND THE LATENT VALUES OF VIRTUAL COMMUNITY NETWORKS By Austin Michael Hooks Heather Palmer Rik Hunter Associate Professor of English Associate Professor of English (Chair) (Committee Member) Matthew Guy Associate Professor of English (Committee Member) CANCEL CULTURE: POSTHUMAN HAUNTOLOGIES IN DIGITAL RHETORIC AND THE LATENT VALUES OF VIRTUAL COMMUNITY NETWORKS By Austin Michael Hooks A Thesis Submitted to the Faculty of the University of Tennessee at Chattanooga in Partial Fulfillment of the Requirements of the Degree of Master of English The University of Tennessee at Chattanooga Chattanooga, Tennessee August 2020 ii Copyright © 2020 By Austin Michael Hooks All Rights Reserved iii ABSTRACT This study explores how modern epideictic practices enact latent community values by analyzing modern call-out culture, a form of public shaming that aims to hold individuals responsible for perceived politically incorrect behavior via social media, and cancel culture, a boycott of such behavior and a variant of call-out culture. As a result, this thesis is mainly concerned with the capacity of words, iterated within the archive of social media, to haunt us— both culturally and informatically. Through hauntology, this study hopes to understand a modern discourse community that is bound by an epideictic framework that specializes in the deconstruction of the individual’s ethos via the constant demonization and incitement of past, current, and possible social media expressions. The primary goal of this study is to understand how these practices function within a capitalistic framework and mirror the performativity of capital by reducing affective human interactions to that of a transaction.
    [Show full text]
  • An Examination of the Impact of Astroturfing on Nationalism: A
    social sciences $€ £ ¥ Article An Examination of the Impact of Astroturfing on Nationalism: A Persuasion Knowledge Perspective Kenneth M. Henrie 1,* and Christian Gilde 2 1 Stiller School of Business, Champlain College, Burlington, VT 05401, USA 2 Department of Business and Technology, University of Montana Western, Dillon, MT 59725, USA; [email protected] * Correspondence: [email protected]; Tel.: +1-802-865-8446 Received: 7 December 2018; Accepted: 23 January 2019; Published: 28 January 2019 Abstract: One communication approach that lately has become more common is astroturfing, which has been more prominent since the proliferation of social media platforms. In this context, astroturfing is a fake grass-roots political campaign that aims to manipulate a certain audience. This exploratory research examined how effective astroturfing is in mitigating citizens’ natural defenses against politically persuasive messages. An experimental method was used to examine the persuasiveness of social media messages related to coal energy in their ability to persuade citizens’, and increase their level of nationalism. The results suggest that citizens are more likely to be persuaded by an astroturfed message than people who are exposed to a non-astroturfed message, regardless of their political leanings. However, the messages were not successful in altering an individual’s nationalistic views at the moment of exposure. The authors discuss these findings and propose how in a long-term context, astroturfing is a dangerous addition to persuasive communication. Keywords: astroturfing; persuasion knowledge; nationalism; coal energy; social media 1. Introduction Astroturfing is the simulation of a political campaign which is intended to manipulate the target audience. Often, the purpose of such a campaign is not clearly recognizable because a political effort is disguised as a grassroots operation.
    [Show full text]
  • Understanding the Challenges of Collaborative Evidence-Based
    Do you have a source for that? Understanding the Challenges of Collaborative Evidence-based Journalism Sheila O’Riordan Gaye Kiely Bill Emerson Joseph Feller Business Information Business Information Business Information Business Information Systems, University Systems, University Systems, University Systems, University College Cork, Ireland College Cork, Ireland College Cork, Ireland College Cork, Ireland [email protected] [email protected] [email protected] [email protected] ABSTRACT consumption has evolved [25]. There has been a move WikiTribune is a pilot news service, where evidence-based towards digital news with increasing user involvement as articles are co-created by professional journalists and a well as the use of social media platforms for accessing and community of volunteers using an open and collaborative discussing current affairs [14,39,43]. As such, the boundaries digital platform. The WikiTribune project is set within an are shifting between professional and amateur contributions. evolving and dynamic media landscape, operating under Traditional news organizations are adding interactive principles of openness and transparency. It combines a features as participatory journalism practices rise (see [8,42]) commercial for-profit business model with an open and the technologies that allow citizens to interact en masse collaborative mode of production with contributions from provide new avenues for engaging in democratic both paid professionals and unpaid volunteers. This deliberation [19]; the “process of reaching reasoned descriptive case study captures the first 12-months of agreement among free and equal citizens” [6:322]. WikiTribune’s operations to understand the challenges and opportunities within this hybrid model of production. We use With these changes, a number of challenges have arisen.
    [Show full text]
  • Political Astroturfing Across the World
    Political astroturfing across the world Franziska B. Keller∗ David Schochy Sebastian Stierz JungHwan Yang§ Paper prepared for the Harvard Disinformation Workshop Update 1 Introduction At the very latest since the Russian Internet Research Agency’s (IRA) intervention in the U.S. presidential election, scholars and the broader public have become wary of coordi- nated disinformation campaigns. These hidden activities aim to deteriorate the public’s trust in electoral institutions or the government’s legitimacy, and can exacerbate political polarization. But unfortunately, academic and public debates on the topic are haunted by conceptual ambiguities, and rely on few memorable examples, epitomized by the often cited “social bots” who are accused of having tried to influence public opinion in various contemporary political events. In this paper, we examine “political astroturfing,” a particular form of centrally co- ordinated disinformation campaigns in which participants pretend to be ordinary citizens acting independently (Kovic, Rauchfleisch, Sele, & Caspar, 2018). While the accounts as- sociated with a disinformation campaign may not necessarily spread incorrect information, they deceive the audience about the identity and motivation of the person manning the ac- count. And even though social bots have been in the focus of most academic research (Fer- rara, Varol, Davis, Menczer, & Flammini, 2016; Stella, Ferrara, & De Domenico, 2018), seemingly automated accounts make up only a small part – if any – of most astroturf- ing campaigns. For instigators of astroturfing, relying exclusively on social bots is not a promising strategy, as humans are good at detecting low-quality information (Pennycook & Rand, 2019). On the other hand, many bots detected by state-of-the-art social bot de- tection methods are not part of an astroturfing campaign, but unrelated non-political spam ∗Hong Kong University of Science and Technology yThe University of Manchester zGESIS – Leibniz Institute for the Social Sciences §University of Illinois at Urbana-Champaign 1 bots.
    [Show full text]
  • Trump's Disinformation 'Magaphone'. Consequences, First Lessons and Outlook
    BRIEFING Trump's disinformation 'magaphone' Consequences, first lessons and outlook SUMMARY The deadly insurrection at the US Capitol on 6 January 2021 was a significant cautionary example of the offline effects of online disinformation and conspiracy theories. The historic democratic crisis this has sparked − adding to a number of other historic crises the US is currently battling − provides valuable lessons not only for the United States, but also for Europe and the democratic world. The US presidential election and its aftermath saw domestic disinformation emerging as a more visible immediate threat than disinformation by third countries. While political violence has been the most tangible physical effect of manipulative information, corrosive conspiracy theories have moved from the fringes to the heart of political debate, normalising extremist rhetoric. At the same time, recent developments have confirmed that the lines between domestic and foreign attempts to undermine democracy are increasingly blurred. While the perceived weaknesses in democratic systems are − unsurprisingly − celebrated as a victory for authoritarian state actors, links between foreign interference and domestic terrorism are under growing scrutiny. The question of how to depolarise US society − one of a long list of challenges facing the Biden Administration − is tied to the polarised media environment. The crackdown by major social media platforms on Donald Trump and his supporters has prompted far-right groups to abandon the established information ecosystem to join right-wing social media. This could further accelerate the ongoing fragmentation of the US infosphere, cementing the trend towards separate realities. Ahead of the proposed Democracy Summit − a key objective of the Biden Administration − tempering the 'sword of democracy' has risen to the top of the agenda on both sides of the Atlantic.
    [Show full text]
  • Researching Violent Extremism the State of Play
    Researching Violent Extremism The State of Play J.M. Berger RESOLVE NETWORK | JUNE 2019 Researching Violent Extremism Series The views in this publication are those of the author. They do not necessarily reflect the views of the RESOLVE Network, its partners, the U.S. Institute of Peace, or any U.S. government agency. 2 RESOLVE NETWORK RESEARCH REPORT NO. 1 | LAKE CHAD BASIN RESEARCH SERIES The study of violent extremism is entering a new phase, with shifts in academic focus and policy direction, as well as a host of new and continuing practical and ethical challenges. This chapter will present an overview of the challenges ahead and discuss some strategies for improving the state of research. INTRODUCTION The field of terrorism studies has vastly expanded over the last two decades. As an illustrative example, the term “terrorism” has been mentioned in an average of more than 60,000 Google Scholar results per year since 2010 alone, including academic publications and cited works. While Google Scholar is an admittedly imprecise tool, the index provides some insights on relative trends in research on terrorism. While almost 5,000 publications indexed on Google Scholar address, to a greater or lesser extent, the question of “root causes of terrorism, at the beginning of this marathon of output, which started soon after the terrorist attacks of September 11, 2001, only a fractional number of indexed publications addressed “extremism.” Given the nature of terrorist movements, however, this should be no great mys- tery. The root cause of terrorism is extremism. Figure 1: Google Scholar Results per Year1 Publications mentioning terrorism Publications mentioning extremism 90,000 80,000 70,000 60,000 50,000 40,000 30,000 20,000 Google Scholar Results 10,000 0 Year 1 Google Scholar, accessed May 30, 2019.
    [Show full text]
  • Articles & Reports
    1 Reading & Resource List on Information Literacy Articles & Reports Adegoke, Yemisi. "Like. Share. Kill.: Nigerian police say false information on Facebook is killing people." BBC News. Accessed November 21, 2018. https://www.bbc.co.uk/news/resources/idt- sh/nigeria_fake_news. See how Facebook posts are fueling ethnic violence. ALA Public Programs Office. “News: Fake News: A Library Resource Round-Up.” American Library Association. February 23, 2017. http://www.programminglibrarian.org/articles/fake-news-library-round. ALA Public Programs Office. “Post-Truth: Fake News and a New Era of Information Literacy.” American Library Association. Accessed March 2, 2017. http://www.programminglibrarian.org/learn/post-truth- fake-news-and-new-era-information-literacy. This has a 45-minute webinar by Dr. Nicole A. Cook, University of Illinois School of Information Sciences, which is intended for librarians but is an excellent introduction to fake news. Albright, Jonathan. “The Micro-Propaganda Machine.” Medium. November 4, 2018. https://medium.com/s/the-micro-propaganda-machine/. In a three-part series, Albright critically examines the role of Facebook in spreading lies and propaganda. Allen, Mike. “Machine learning can’g flag false news, new studies show.” Axios. October 15, 2019. ios.com/machine-learning-cant-flag-false-news-55aeb82e-bcbb-4d5c-bfda-1af84c77003b.html. Allsop, Jon. "After 10,000 'false or misleading claims,' are we any better at calling out Trump's lies?" Columbia Journalism Review. April 30, 2019. https://www.cjr.org/the_media_today/trump_fact- check_washington_post.php. Allsop, Jon. “Our polluted information ecosystem.” Columbia Journalism Review. December 11, 2019. https://www.cjr.org/the_media_today/cjr_disinformation_conference.php. Amazeen, Michelle A.
    [Show full text]
  • Wikipedia's Jimmy Wales Has Launched an Alternative to Facebook and Twitter 11/18/19, 8:11 PM
    Wikipedia's Jimmy Wales has launched an alternative to Facebook and Twitter 11/18/19, 8:11 PM Wikipedia's Jimmy Wales has launched an alternative to Facebook and Twitter In a nutshell: Hardly anyone would say that social media is good for you, yet billions use it in one form or another (often concurrently) every single day. These platforms make billions of dollars by addicting users and getting them to click ads. One Wikipedia co-founder wants to change that with a new social media site that is supported by the users rather than big advertisers. Wikipedia co-founder Jimmy Wales is launching a social-media website called WT: Social. The platform aims to compete with Facebook and Twitter, except instead of funding it using advertising, Wales is taking a page from the Wikipedia playbook and financing it through user donations. "The business model of social media companies, of pure advertising, is problematic," Wales told Financial Times. "It turns out the huge winner is low-quality content." WT: Social got its start as Wikitribune, a site that published original news stories with the community fact-checking and sub-editing articles. The venture never gained much traction, so Wales is moving it to the new platform with a more social networking focus. "Instead of optimizing our algorithm to addict you and keep you clicking, we will only make money if you voluntarily choose to support us – which means that our goal is not clicks but actually being meaningful to your life." The site will still post articles, but instead of giving priority to content with the most "Likes," its algorithms will list the newest stories first.
    [Show full text]
  • What Is Astroturfing? Churchofgodbigsandy.Com
    What Is Astroturfing? The following information is from the “Edifying the Body” section of the Church of God Big Sandy’s website, churchofgodbigsandy.com. It was posted for the weekend of Dec. 10, 2016. Compiled by Dave Havir A definition titled “Astroturf” was posted at powerbase.info. Following is an excerpt of the definition. __________ “Astroturf” refers to grassroots groups or coalitions which are actually fake; often created or heavily funded by corporations, public relations firms, indus- try trade associations, and political interests. Astroturfing is used by organizations to give the illusion of genuine public support to their cause, manufacturing public opinion in what some commen- tators have called “democracy for hire.” As a deceptive use of third-party technique, Astroturfing can be considered a form of propaganda. Unlike genuine grassroots activism, which tends to be people-rich but cash- poor, Astroturf activism is normally people-poor but cash-rich. Astroturf campaigns work by recruiting the support of less-informed activists and individuals to their cause, often by means of deception. Astroturfing can mislead the public into believing that the views of the astro- turfer are mainstream and that widespread genuine support actually exists, when in most cases it does not. Deceptive Astroturf campaigns are thus most likely to occur where the interests of wealthy or powerful interests come into conflict with the interests of the public. ★★★★★ A definition titled “Astroturfing” was posted at weebly.com. Following is an excerpt of the definition. __________ “Astroturfing” denotes political, advertising, or public relations campaigns that are formally planned by an organization, but are disguised as sponta- neous, popular “grassroots” behavior.
    [Show full text]
  • Response to Evan Williams–
    Email Interview on June 22, 2017 with Geert Lovink by Davide Nitrosi, Italian journalist with a group of daily papers such as Bologna’s Resto del Carlino, Florence’s Nazione and Milan’s Giorno. Davide Nitrosi: Once Twitter was viewed as a tool of liberation. During the Arab Spring uprisings in the Middle East and the Iran riots before that Twitter was the media of freedom. We thought Internet was against tyranny. Now we see the abuse of social networks. Twitter founder, Evan Williams, said to the New York Times that Twitter’s role in the Trump election was a very bad thing. What’s your response to this article? Geert Lovink: The phrase “The Internet is broken” than Evan has in fact circulated amongst geeks and engineers since 2013/14 in response to the revelations of Edward Snowden (the campaign used to have the http://internetisbroken.org/ website). The slogan refers to the widely felt loss of privacy, caused by what we now call ‘surveillance capitalism’, in which the industrial-surveillance apparatus around the NSA and affiliated secret services broke the collective dream of a public internet in which users were in charge. The inventor of the World Wide Web, Tim-Berners Lee, is playing an important role here, in the background. Or look at Wikitribune, crowdsourcing evidence-based journalism, the latest intiative of Jimmy Wales (Wikipedia). This issue has been a long one in the making. In that sense, the response of EV comes late. We’re now more than half year into the ‘fake news’ meme (that comes from elsewhere anyway, for instance Ukraine).
    [Show full text]
  • Great Meme War:” the Alt-Right and Its Multifarious Enemies
    Angles New Perspectives on the Anglophone World 10 | 2020 Creating the Enemy The “Great Meme War:” the Alt-Right and its Multifarious Enemies Maxime Dafaure Electronic version URL: http://journals.openedition.org/angles/369 ISSN: 2274-2042 Publisher Société des Anglicistes de l'Enseignement Supérieur Electronic reference Maxime Dafaure, « The “Great Meme War:” the Alt-Right and its Multifarious Enemies », Angles [Online], 10 | 2020, Online since 01 April 2020, connection on 28 July 2020. URL : http:// journals.openedition.org/angles/369 This text was automatically generated on 28 July 2020. Angles. New Perspectives on the Anglophone World is licensed under a Creative Commons Attribution- NonCommercial-ShareAlike 4.0 International License. The “Great Meme War:” the Alt-Right and its Multifarious Enemies 1 The “Great Meme War:” the Alt- Right and its Multifarious Enemies Maxime Dafaure Memes and the metapolitics of the alt-right 1 The alt-right has been a major actor of the online culture wars of the past few years. Since it came to prominence during the 2014 Gamergate controversy,1 this loosely- defined, puzzling movement has achieved mainstream recognition and has been the subject of discussion by journalists and scholars alike. Although the movement is notoriously difficult to define, a few overarching themes can be delineated: unequivocal rejections of immigration and multiculturalism among most, if not all, alt- right subgroups; an intense criticism of feminism, in particular within the manosphere community, which itself is divided into several clans with different goals and subcultures (men’s rights activists, Men Going Their Own Way, pick-up artists, incels).2 Demographically speaking, an overwhelming majority of alt-righters are white heterosexual males, one of the major social categories who feel dispossessed and resentful, as pointed out as early as in the mid-20th century by Daniel Bell, and more recently by Michael Kimmel (Angry White Men 2013) and Dick Howard (Les Ombres de l’Amérique 2017).
    [Show full text]
  • Does 'Deplatforming' Work to Curb Hate Speech and Calls for Violence?
    Exercise Nº Professor´s Name Mark 1. Reading Comp. ………………… .…/20 2. Paraphrasing ………………… .…/30 Total Part I (Min. 26).…/50 Part Part I 3. Essay ………………… …/50 Re correction …………………… …/50 Essay Final Mark …………………… …/50 Part Part II (do NOT fill in) Total Part II (Min.26) …/50 CARRERA DE TRADUCTOR PÚBLICO - ENTRANCE EXAMINATION – FEBRERO 2021 NOMBRE y APELLIDO: …………………………………………………………………………… Nº de ORDEN: (NO es el DNI) ……………………………………………………………………. Please read the text carefully and then choose the best answer. Remember the questions do not follow the order of the reading passage (Paragraphs are numbered for the sake of correction) Does ‘deplatforming’ work to curb hate speech and calls for violence? 3 experts in online communications weigh in 1- In the wake of the assault on the U.S. Capitol on Jan. 6, Twitter permanently suspended Donald Trump’s personal account, and Google, Apple, and Amazon shunned Parler, which at least temporarily shut down the social media platform favored by the far right. 2- Dubbed “deplatforming,” these actions restrict the ability of individuals and communities to communicate with each other and the public. Deplatforming raises ethical and legal questions, but foremost is the question of whether it is an effective strategy to reduce hate speech and calls for violence on social media. 3- The Conversation U.S. asked three experts in online communications whether deplatforming works and what happens when technology companies attempt it. Jeremy Blackburn, assistant professor of computer science, Binghamton University 4-Does deplatforming work from a technical perspective? Gab was the first “major” platform subject to deplatforming efforts, first with removal from app stores and, after the Tree of Life shooting, the withdrawal of cloud infrastructure providers, domain name providers and other Web-related services.
    [Show full text]