Trends in Online Influence Efforts
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
How Russia Tried to Start a Race War in the United States
Michigan Journal of Race and Law Volume 24 2019 Virtual Hatred: How Russia Tried to Start a Race War in the United States William J. Aceves California Western School of Law Follow this and additional works at: https://repository.law.umich.edu/mjrl Part of the Communications Law Commons, Internet Law Commons, and the Law and Race Commons Recommended Citation William J. Aceves, Virtual Hatred: How Russia Tried to Start a Race War in the United States, 24 MICH. J. RACE & L. 177 (2019). Available at: https://repository.law.umich.edu/mjrl/vol24/iss2/2 This Article is brought to you for free and open access by the Journals at University of Michigan Law School Scholarship Repository. It has been accepted for inclusion in Michigan Journal of Race and Law by an authorized editor of University of Michigan Law School Scholarship Repository. For more information, please contact [email protected]. VIRTUAL HATRED: HOW RUSSIA TRIED TO START A RACE WAR in the UNITED STATES William J. Aceves* During the 2016 U.S. presidential election, the Russian government engaged in a sophisticated strategy to influence the U.S. political system and manipulate American democracy. While most news reports have focused on the cyber-attacks aimed at Democratic Party leaders and possible contacts between Russian officials and the Trump presidential campaign, a more pernicious intervention took place. Throughout the campaign, Russian operatives created hundreds of fake personas on social media platforms and then posted thousands of advertisements and messages that sought to promote racial divisions in the United States. This was a coordinated propaganda effort. -
Automated Tackling of Disinformation
Automated tackling of disinformation STUDY Panel for the Future of Science and Technology European Science-Media Hub EPRS | European Parliamentary Research Service Scientific Foresight Unit (STOA) PE 624.278 – March 2019 EN Automated tackling of disinformation Major challenges ahead This study maps and analyses current and future threats from online misinformation, alongside currently adopted socio-technical and legal approaches. The challenges of evaluating their effectiveness and practical adoption are also discussed. Drawing on and complementing existing literature, the study summarises and analyses the findings of relevant journalistic and scientific studies and policy reports in relation to detecting, containing and countering online disinformation and propaganda campaigns. It traces recent developments and trends and identifies significant new or emerging challenges. It also addresses potential policy implications for the EU of current socio-technical solutions. ESMH | European Science-Media Hub AUTHORS This study was written by Alexandre Alaphilippe, Alexis Gizikis and Clara Hanot of EU DisinfoLab, and Kalina Bontcheva of The University of Sheffield, at the request of the Panel for the Future of Science and Technology (STOA). It has been financed under the European Science and Media Hub budget and managed by the Scientific Foresight Unit within the Directorate-General for Parliamentary Research Services (EPRS) of the Secretariat of the European Parliament. Acknowledgements The authors wish to thank all respondents to the online survey, as well as first draft, WeVerify, InVID, PHEME, REVEAL, and all other initiatives that contributed materials to the study. ADMINISTRATOR RESPONSIBLE Mihalis Kritikos, Scientific Foresight Unit To contact the publisher, please e-mail [email protected] LINGUISTIC VERSION Original: EN Manuscript completed in March 2019. -
ASD-Covert-Foreign-Money.Pdf
overt C Foreign Covert Money Financial loopholes exploited by AUGUST 2020 authoritarians to fund political interference in democracies AUTHORS: Josh Rudolph and Thomas Morley © 2020 The Alliance for Securing Democracy Please direct inquiries to The Alliance for Securing Democracy at The German Marshall Fund of the United States 1700 18th Street, NW Washington, DC 20009 T 1 202 683 2650 E [email protected] This publication can be downloaded for free at https://securingdemocracy.gmfus.org/covert-foreign-money/. The views expressed in GMF publications and commentary are the views of the authors alone. Cover and map design: Kenny Nguyen Formatting design: Rachael Worthington Alliance for Securing Democracy The Alliance for Securing Democracy (ASD), a bipartisan initiative housed at the German Marshall Fund of the United States, develops comprehensive strategies to deter, defend against, and raise the costs on authoritarian efforts to undermine and interfere in democratic institutions. ASD brings together experts on disinformation, malign finance, emerging technologies, elections integrity, economic coercion, and cybersecurity, as well as regional experts, to collaborate across traditional stovepipes and develop cross-cutting frame- works. Authors Josh Rudolph Fellow for Malign Finance Thomas Morley Research Assistant Contents Executive Summary �������������������������������������������������������������������������������������������������������������������� 1 Introduction and Methodology �������������������������������������������������������������������������������������������������� -
Taming the Trolls: the Need for an International Legal Framework to Regulate State Use of Disinformation on Social Media
Taming the Trolls: The Need for an International Legal Framework to Regulate State Use of Disinformation on Social Media * ASHLEY C. NICOLAS INTRODUCTION Consider a hypothetical scenario in which hundreds of agents of the Russian GRU arrive in the United States months prior to a presidential election.1 The Russian agents spend the weeks leading up to the election going door to door in vulnerable communities, spreading false stories intended to manipulate the population into electing a candidate with policies favorable to Russian positions. The agents set up television stations, use radio broadcasts, and usurp the resources of local newspapers to expand their reach and propagate falsehoods. The presence of GRU agents on U.S. soil is an incursion into territorial integrity⎯a clear invasion of sovereignty.2 At every step, Russia would be required to expend tremendous resources, overcome traditional media barriers, and risk exposure, making this hypothetical grossly unrealistic. Compare the hypothetical with the actual actions of the Russians during the 2016 U.S. presidential election. Sitting behind computers in St. Petersburg, without ever setting foot in the United States, Russian agents were able to manipulate the U.S. population in the most sacred of domestic affairs⎯an election. Russian “trolls” targeted vulnerable populations through social media, reaching millions of users at a minimal cost and without reliance on established media institutions.3 Without using * Georgetown Law, J.D. expected 2019; United States Military Academy, B.S. 2009; Loyola Marymount University M.Ed. 2016. © 2018, Ashley C. Nicolas. The author is a former U.S. Army Intelligence Officer. -
Digital Populism: Trolls and Political Polarization of Twitter in Turkey
International Journal of Communication 11(2017), 4093–4117 1932–8036/20170005 Digital Populism: Trolls and Political Polarization of Twitter in Turkey ERGİN BULUT Koç University, Turkey ERDEM YÖRÜK Koç University, Turkey University of Oxford, UK This article analyzes political trolling in Turkey through the lens of mediated populism. Twitter trolling in Turkey has diverged from its original uses (i.e., poking fun, flaming, etc.) toward government-led polarization and right-wing populism. Failing to develop an effective strategy to mobilize online masses, Turkey’s ruling Justice and Development Party (JDP/AKP) relied on the polarizing performances of a large progovernment troll army. Trolls deploy three features of JDP’s populism: serving the people, fetish of the will of the people, and demonization. Whereas trolls traditionally target and mock institutions, Turkey’s political trolls act on behalf of the establishment. They produce a digital culture of lynching and censorship. Trolls’ language also impacts pro-JDP journalists who act like trolls and attack journalists, academics, and artists critical of the government. Keywords: trolls, mediated populism, Turkey, political polarization, Twitter Turkish media has undergone a transformation during the uninterrupted tenure of the ruling Justice and Development Party (JDP) since 2002. Not supported by the mainstream media when it first came to power, JDP created its own media army and transformed the mainstream media’s ideological composition. What has, however, destabilized the entire media environment was the Gezi Park protests of summer 2013.1 Activists’ use of social media not only facilitated political organizing, but also turned the news environment upside down. Having recognized that the mainstream media was not trustworthy, oppositional groups migrated to social media for organizing and producing content. -
Recent Trends in Online Foreign Influence Efforts
Recent Trends in Online Foreign Influence Efforts Diego A. Martin, Jacob N. Shapiro, Michelle Nedashkovskaya Woodrow Wilson School of Public and International Affairs Princeton University Princeton, New Jersey, United States E-mail: [email protected] Email: [email protected] E-mail: [email protected] Abstract: Foreign governments have used social media to influence politics in a range of countries by promoting propaganda, advocating controversial viewpoints, and spreading disinformation. We analyze 53 distinct foreign influence efforts (FIEs) targeting 24 different countries from 2013 through 2018. FIEs are defined as (i) coordinated campaigns by one state to impact one or more specific aspects of politics in another state (ii) through media channels, including social media, (iii) by producing content designed to appear indigenous to the target state. The objective of such campaigns can be quite broad and to date have included influencing political decisions by shaping election outcomes at various levels, shifting the political agenda on topics ranging from health to security, and encouraging political polarization. We draw on more than 460 media reports to identify FIEs, track their progress, and classify their features. Introduction Information and Communications Technologies (ICTs) have changed the way people communi- cate about politics and access information on a wide range of topics (Foley 2004, Chigona et al. 2009). Social media in particular has transformed communication between leaders and voters by enabling direct politician-to-voter engagement outside traditional avenues, such as speeches and press conferences (Ott 2017). In the 2016 U.S. presidential election, for example, social media platforms were more widely viewed than traditional editorial media and were central to the campaigns of both Democratic candidate Hillary Clinton and Republican candidate Donald Trump (Enli 2017). -
VYTAUTAS MAGNUS UNIVERSITY Maryna Kupriienko INFLUENCING
VYTAUTAS MAGNUS UNIVERSITY FACULTY OF POLITICAL SCIENCE AND DIPLOMACY PUBLIC COMMUNICATIONS DEPARTMENT Maryna Kupriienko INFLUENCING PUBLIC OPINION IN SOCIAL MEDIA. TROLLING AND DISINFORMATION STRATEGIES Final Master Thesis Journalism and Media Industries Study Program, state code 621P50002 Degree in Journalism Supervisor prof. Robert van Voren (acad. title, name, surname) Defended prof. Šarūnas Liekis (Dean of the Faculty) Kaunas, 2019 VYTAUTO DIDŽIOJO UNIVERSITETAS POLITIKOS MOKSLŲ IR DIPLOMATIJOS FAKULTETAS VIEŠOSIOS KOMUNIKACIJOS KATEDRA Maryna Kupriienko VIEŠOS NUOMONĖS ĮTAKA SOCIALINĖSE MEDIJOSE. TROLINIMO IR DEZINFORMACIJOS STRATEGIJOS Magistro baigiamasis darbas Žurnalistikos ir medijų industrijų studijų programa, valstybinis kodas 621P50002 Žurnalistikos studijų kryptis Vadovas (-ė) prof. Robert van Voren (Moksl. laipsnis, vardas, pavardė) Apginta prof. Šarūnas Liekis (Fakulteto/studijų instituto dekanas/direktorius) Kaunas, 2019 CONTENTS SUMMARY ................................................................................................................................................ 4 SANTRAUKA ............................................................................................................................................ 5 INTRODUCTION ..................................................................................................................................... 6 1. THEORETICAL FRAMEWORK ........................................................................................................ 9 1.1. Theoretical -
COVID-19 Effects and Russian Disinformation Campaigns
Homeland Security Affairs Journal SPECIAL COVID-19 ISSUE COVID-19 Effects and Russian Disinformation Campaigns By Wesley R. Moy and Kacper Gradon 2 COVID-19 Effects and Russian Disinformation Campaigns By Wesley R. Moy and Kacper Gradon Abstract The effects of the novel coronavirus and its related disease COVID-19 have been far reaching, touching American society and its Western partners’ health and mortality, economics, social relationships, and politics. The effects have highlighted longstanding societal divisions, disproportionately harming minority groups and the socioeconomically disadvantaged with Black Americans and their communities hit especially hard. The problem being considered is the potential for Russian malign foreign influence to exploit the divides exacerbated by the coronavirus and target the United States and its European allies during the period leading to the 2020 elections and beyond. Possible coronavirus effects are inventoried and categorized into economic, healthcare, and environmental issue areas that may be leveraged and weaponized. The article includes the scenarios of such weaponization, including the description of focal points that may be attacked. We conclude with a set of recommendations to counter malign foreign influence. Suggested Citation Moy, Wesley and Kacper Gradon. “COVID-19 Effects and Russian Disinformation” Homeland Security Affairs 16, Article 8 (December, 2020) www.hsaj.org/articles16533. Introduction In late-May 2020, the novel coronavirus (SARS-CoV-2) and its related illness COVID-19, had resulted in the deaths of more than 100,000 Americans and infected nearly 1.7 million. By mid- October 2020, deaths had increased to over 210,000 and nearly 8,000,000 have been infected. -
Empirical Comparative Analysis of Bot Evidence in Social Networks
Bots in Nets: Empirical Comparative Analysis of Bot Evidence in Social Networks Ross Schuchard1(&) , Andrew Crooks1,2 , Anthony Stefanidis2,3 , and Arie Croitoru2 1 Computational Social Science Program, Department of Computational and Data Sciences, George Mason University, Fairfax, VA 22030, USA [email protected] 2 Department of Geography and Geoinformation Science, George Mason University, Fairfax, VA 22030, USA 3 Criminal Investigations and Network Analysis Center, George Mason University, Fairfax, VA 22030, USA Abstract. The emergence of social bots within online social networks (OSNs) to diffuse information at scale has given rise to many efforts to detect them. While methodologies employed to detect the evolving sophistication of bots continue to improve, much work can be done to characterize the impact of bots on communication networks. In this study, we present a framework to describe the pervasiveness and relative importance of participants recognized as bots in various OSN conversations. Specifically, we harvested over 30 million tweets from three major global events in 2016 (the U.S. Presidential Election, the Ukrainian Conflict and Turkish Political Censorship) and compared the con- versational patterns of bots and humans within each event. We further examined the social network structure of each conversation to determine if bots exhibited any particular network influence, while also determining bot participation in key emergent network communities. The results showed that although participants recognized as social bots comprised only 0.28% of all OSN users in this study, they accounted for a significantly large portion of prominent centrality rankings across the three conversations. This includes the identification of individual bots as top-10 influencer nodes out of a total corpus consisting of more than 2.8 million nodes. -
Disinformation, and Influence Campaigns on Twitter 'Fake News'
Disinformation, ‘Fake News’ and Influence Campaigns on Twitter OCTOBER 2018 Matthew Hindman Vlad Barash George Washington University Graphika Contents Executive Summary . 3 Introduction . 7 A Problem Both Old and New . 9 Defining Fake News Outlets . 13 Bots, Trolls and ‘Cyborgs’ on Twitter . 16 Map Methodology . 19 Election Data and Maps . 22 Election Core Map Election Periphery Map Postelection Map Fake Accounts From Russia’s Most Prominent Troll Farm . 33 Disinformation Campaigns on Twitter: Chronotopes . 34 #NoDAPL #WikiLeaks #SpiritCooking #SyriaHoax #SethRich Conclusion . 43 Bibliography . 45 Notes . 55 2 EXECUTIVE SUMMARY This study is one of the largest analyses to date on how fake news spread on Twitter both during and after the 2016 election campaign. Using tools and mapping methods from Graphika, a social media intelligence firm, we study more than 10 million tweets from 700,000 Twitter accounts that linked to more than 600 fake and conspiracy news outlets. Crucially, we study fake and con- spiracy news both before and after the election, allowing us to measure how the fake news ecosystem has evolved since November 2016. Much fake news and disinformation is still being spread on Twitter. Consistent with other research, we find more than 6.6 million tweets linking to fake and conspiracy news publishers in the month before the 2016 election. Yet disinformation continues to be a substantial problem postelection, with 4.0 million tweets linking to fake and conspiracy news publishers found in a 30-day period from mid-March to mid-April 2017. Contrary to claims that fake news is a game of “whack-a-mole,” more than 80 percent of the disinformation accounts in our election maps are still active as this report goes to press. -
Hacks, Leaks and Disruptions | Russian Cyber Strategies
CHAILLOT PAPER Nº 148 — October 2018 Hacks, leaks and disruptions Russian cyber strategies EDITED BY Nicu Popescu and Stanislav Secrieru WITH CONTRIBUTIONS FROM Siim Alatalu, Irina Borogan, Elena Chernenko, Sven Herpig, Oscar Jonsson, Xymena Kurowska, Jarno Limnell, Patryk Pawlak, Piret Pernik, Thomas Reinhold, Anatoly Reshetnikov, Andrei Soldatov and Jean-Baptiste Jeangène Vilmer Chaillot Papers HACKS, LEAKS AND DISRUPTIONS RUSSIAN CYBER STRATEGIES Edited by Nicu Popescu and Stanislav Secrieru CHAILLOT PAPERS October 2018 148 Disclaimer The views expressed in this Chaillot Paper are solely those of the authors and do not necessarily reflect the views of the Institute or of the European Union. European Union Institute for Security Studies Paris Director: Gustav Lindstrom © EU Institute for Security Studies, 2018. Reproduction is authorised, provided prior permission is sought from the Institute and the source is acknowledged, save where otherwise stated. Contents Executive summary 5 Introduction: Russia’s cyber prowess – where, how and what for? 9 Nicu Popescu and Stanislav Secrieru Russia’s cyber posture Russia’s approach to cyber: the best defence is a good offence 15 1 Andrei Soldatov and Irina Borogan Russia’s trolling complex at home and abroad 25 2 Xymena Kurowska and Anatoly Reshetnikov Spotting the bear: credible attribution and Russian 3 operations in cyberspace 33 Sven Herpig and Thomas Reinhold Russia’s cyber diplomacy 43 4 Elena Chernenko Case studies of Russian cyberattacks The early days of cyberattacks: 5 the cases of Estonia, -
Artificial Intelligence and Countering Violent Extremism: a Primer
Artificial Intelligence and Countering Violent Extremism: A Primer ByMarie Hxxnry Schroeter Mxxhm, Inxxs Oxxlxxmxxnxx, xxnd Fxxnxx Sxxngxxr GNET xxsis a xxspecial spxxcxxl project prxxjxxct delivered dxxlxxvxxrxxd by the International by thxx Intxxrnxxtxxnxxl Centre Cxxntrxx fxxrfor the thxx Study Stxxdy of Radicalisation,xxf Rxxdxxcxxlxxsxxtxxn, King’s College Kxxng’s London. Cxxllxxgxx Lxxndxxn. The author of this report is Marie Schroeter, Mercator Fellow on New Technology in International Relations: Potentials and Limitations of Artifical Intelligence to Prevent Violent Extremism Online The Global Network on Extremism and Technology (GNET) is an academic research initiative backed by the Global Internet Forum to Counter Terrorism (GIFCT), an independent but industry‑funded initiative for better understanding, and counteracting, terrorist use of technology. GNET is convened and led by the International Centre for the Study of Radicalisation (ICSR), an academic research centre based within the Department of War Studies at King’s College London. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing those, either expressed or implied, of GIFCT, GNET or ICSR. We would like to thank Tech Against Terrorism for their support with this report. CONTACT DETAILS For questions, queries and additional copies of this report, please contact: ICSR King’s College London Strand London WC2R 2LS United Kingdom T. +44 20 7848 2098 E. [email protected] Twitter: @GNET_research Like all other GNET publications, this report can be downloaded free of charge from the GNET website at www.gnet‑research.org. © GNET Artificial Intelligence and Countering Violent Extremism: A Primer Executive Summary Radicalisation can take place offline as well as online.