Modeling and Identifying Russian Troll Accounts on Twitter

Total Page:16

File Type:pdf, Size:1020Kb

Modeling and Identifying Russian Troll Accounts on Twitter Still out there: Modeling and Identifying Russian Troll Accounts on Twitter Jane Im Eshwar Chandrasekharan Jackson Sargent [email protected] [email protected] [email protected] University of Michigan University of Illinois at University of Michigan Ann Arbor, Michigan Urbana-Champaign Ann Arbor, Michigan Urbana, Illinois Paige Lighthammer Taylor Denby Ankit Bhargava [email protected] [email protected] [email protected] University of Michigan University of Michigan University of Michigan Ann Arbor, Michigan Ann Arbor, Michigan Ann Arbor, Michigan Libby Hemphill David Jurgens Eric Gilbert [email protected] [email protected] [email protected] University of Michigan University of Michigan University of Michigan Ann Arbor, Michigan Ann Arbor, Michigan Ann Arbor, Michigan ABSTRACT KEYWORDS There is evidence that Russia’s Internet Research Agency attempted Russian troll, misinformation, social media, political manipulation, to interfere with the 2016 U.S. election by running fake accounts political elections on Twitter—often referred to as “Russian trolls”. In this work, we: ACM Reference Format: 1) develop machine learning models that predict whether a Twitter Jane Im, Eshwar Chandrasekharan, Jackson Sargent, Paige Lighthammer, account is a Russian troll within a set of 170K control accounts; and, Taylor Denby, Ankit Bhargava, Libby Hemphill, David Jurgens, and Eric 2) demonstrate that it is possible to use this model to find active Gilbert. 2020. Still out there: Modeling and Identifying Russian Troll Ac- accounts on Twitter still likely acting on behalf of the Russian counts on Twitter. In 12th ACM Conference on Web Science (WebSci ’20), state. Using both behavioral and linguistic features, we show that July 6–10, 2020, Southampton, United Kingdom. ACM, New York, NY, USA, it is possible to distinguish between a troll and a non-troll with a 10 pages. https://doi.org/10.1145/3394231.3397889 precision of 78.5% and an AUC of 98.9%, under cross-validation. Applying the model to out-of-sample accounts still active today, 1 INTRODUCTION we find that up to 2.6% of top journalists’ mentions are occupied It is widely believed that Russia’s Internet Research Agency (IRA) by Russian trolls. These findings imply that the Russian trolls are tried to interfere with the 2016 U.S. election as well as other elections very likely still active today. Additional analysis shows that they by running fake accounts on Twitter—often called the "Russian are not merely software-controlled bots, and manage their online troll" accounts [13, 16, 34]. This interference could have immense identities in various complex ways. Finally, we argue that if it is consequences considering the viral nature of some tweets [25, 26], possible to discover these accounts using externally-accessible data, the number of users exposed to Russian trolls’ content [19, 33], and then the platforms—with access to a variety of private internal the critical role social media have played in past political campaigns signals—should succeed at similar or better rates. [7]. In this paper, we develop models on a dataset of Russian trolls active on Twitter during the 2016 U.S. elections to predict currently CCS CONCEPTS active Russian trolls. We construct machine learning classifiers using profile elements, behavioral features, language distribution, • Human-centered computing ! Empirical studies in collab- function word usage, and linguistic features, on a highly unbalanced orative and social computing. dataset of Russian troll accounts (2.2K accounts, or 1.4% of our sample) released by Twitter1 and “normal”, control accounts (170K accounts, or 98.6% of our sample) collected by the authors. (See Permission to make digital or hard copies of all or part of this work for personal or Figure 1 for a visual overview of the process used in this work.) Our classroom use is granted without fee provided that copies are not made or distributed goals are to determine whether “new” trolls can be identified by for profit or commercial advantage and that copies bear this notice and the full citation models built on “old” trolls and to demonstrate that troll detection on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or is both possible and efficient, even with “old” data. republish, to post on servers or to redistribute to lists, requires prior specific permission We find that it is possible to disambiguate between a Russian and/or a fee. Request permissions from [email protected]. troll account and a large number of these randomly selected con- WebSci ’20, July 6–10, 2020, Southampton, United Kingdom © 2020 Copyright held by the owner/author(s). Publication rights licensed to ACM. trol accounts among users. One model, a simple logistic regression, ACM ISBN 978-1-4503-7989-2/20/07...$15.00 https://doi.org/10.1145/3394231.3397889 1https://about.twitter.com/en_us/values/elections-integrity.html#data WebSci ’20, July 6–10, 2020, Southampton, United Kingdom Im et al. Figure 1: Flowchart illustrating the steps of our research pipeline. achieves a precision of 78.5% and an AUC of 98.9%. Next we asked these accounts using externally-accessible data, then the social whether it was possible to use the model trained on past data to platforms—with access to a variety of private, internal signals— unmask Russian trolls currently active on Twitter (see Figure 2 for should succeed at similar or better rates at finding and deactivating an example)? The logistic regression is attractive in this context Russian troll accounts. as its simplicity seems most likely to generalize to out-of-sample data. Toward that end, we apply our classifier to Twitter accounts 2 RELATED WORK that mentioned high-profile journalists in late 2018. We find the First, we review what is known about Russian’s interference in computational model flags 3.7% of them as statistically likely Rus- Western democracies via online campaigns, and then move on to sian trolls and find reasonable agreement between our classifier the emerging work on these 2016 election related Russian trolls and human labelers. themselves. We conclude by discussing work on social bots, and by Our model allows us to estimate the activity of trolls. As a case reviewing theories of online deception that inform the quantitative study, we estimate the activity of suspected Russian troll accounts approaches in this paper. engaging in one type of adversarial campaign: engaging with promi- nent journalists. Since we have no way of truly knowing which of these model-identified accounts are truly Russian trolls—perhaps 2.1 Russia’s Interference on Political only the IRA knows this—we perform a secondary human eval- Campaigns uation in order to establish consensus on whether the model is While state-level online interference in democratic processes is an identifying validly suspicious accounts. Our human evaluation pro- emerging phenomenon, new research documents Russia’s online cess suggests that roughly 70% of these model-flagged accounts—all political manipulation campaigns in countries other than the United of them still currently active on Twitter—are highly likely to be Rus- States. For instance, previous work has shown that a high volume sian trolls. As a result, we estimate that Russian trolls occupy 2.6% of Russian tweets were generated a few days before the voting day of the mentions of high-profile journalists’ mentions. Moreover, we in the case of the 2016 E.U. Referendum (Brexit Referendum), and find that in contrast with some prevailing narratives surrounding then dropped afterwards [16]. Furthermore, it is suspected that the Russian troll program, the model-flagged accounts do not score Russia is behind the MacronLeaks campaign that occurred during highly on the well-known Botometer scale [9], indicating that they the 2017 French presidential elections period [13], as well as the are not simply automated software agents. Catalonian referendum [34]. Finally, we perform an exploratory open coding of the identity deception strategies used by the currently active accounts discov- 2.2 Emerging Work on the 2016 Russian Trolls ered by our model. For instance, some pretend to be an American mother or a middle-aged white man via profile pictures and descrip- While a brand new area of scholarship, emerging work has exam- tions, but their tweet rates are abnormally high, and their tweets ined the datasets of Russian trolls released by Twitter. Researchers revolve solely around political topics. from Clemson University identified five categories of trolls and This paper makes the following contributions, building on an argued the behavior between these categories were radically dif- emerging line of scholarship around the Russian troll accounts ferent [5]. This was especially marked for left- and right-leaning [5, 6, 17, 33, 35, 42]. First, we show that it is possible to separate accounts (the dataset contains both). For instance, the IRA pro- Russian trolls from other accounts in the data previous to 2019, and moted more left-leaning content than right-leaning on Facebook, that this computational model is still accurate on 2019 data. As a while right-leaning Twitter handles received more engagement. corollary, we believe this work establishes that a large number of [33]. Russian troll accounts are likely to be currently active on Twitter. New work has looked at how the Russian troll accounts were Next, we find that accounts flagged by our model as Russian trolls retweeted in the context of the #BlackLivesMatter movement [35] are not merely bots but use diverse ways to build and manage their – a movement targeted by the trolls. The retweets were divided online identities. Finally, we argue that if it is possible to discover among different political perspectives and the trolls took advantage of this division. There is some disagreement about how predictable Still out there: Modeling and Identifying Russian Troll Accounts on Twitter WebSci ’20, July 6–10, 2020, Southampton, United Kingdom the Russian trolls are.
Recommended publications
  • Automated Tackling of Disinformation
    Automated tackling of disinformation STUDY Panel for the Future of Science and Technology European Science-Media Hub EPRS | European Parliamentary Research Service Scientific Foresight Unit (STOA) PE 624.278 – March 2019 EN Automated tackling of disinformation Major challenges ahead This study maps and analyses current and future threats from online misinformation, alongside currently adopted socio-technical and legal approaches. The challenges of evaluating their effectiveness and practical adoption are also discussed. Drawing on and complementing existing literature, the study summarises and analyses the findings of relevant journalistic and scientific studies and policy reports in relation to detecting, containing and countering online disinformation and propaganda campaigns. It traces recent developments and trends and identifies significant new or emerging challenges. It also addresses potential policy implications for the EU of current socio-technical solutions. ESMH | European Science-Media Hub AUTHORS This study was written by Alexandre Alaphilippe, Alexis Gizikis and Clara Hanot of EU DisinfoLab, and Kalina Bontcheva of The University of Sheffield, at the request of the Panel for the Future of Science and Technology (STOA). It has been financed under the European Science and Media Hub budget and managed by the Scientific Foresight Unit within the Directorate-General for Parliamentary Research Services (EPRS) of the Secretariat of the European Parliament. Acknowledgements The authors wish to thank all respondents to the online survey, as well as first draft, WeVerify, InVID, PHEME, REVEAL, and all other initiatives that contributed materials to the study. ADMINISTRATOR RESPONSIBLE Mihalis Kritikos, Scientific Foresight Unit To contact the publisher, please e-mail [email protected] LINGUISTIC VERSION Original: EN Manuscript completed in March 2019.
    [Show full text]
  • Recent Trends in Online Foreign Influence Efforts
    Recent Trends in Online Foreign Influence Efforts Diego A. Martin, Jacob N. Shapiro, Michelle Nedashkovskaya Woodrow Wilson School of Public and International Affairs Princeton University Princeton, New Jersey, United States E-mail: [email protected] Email: [email protected] E-mail: [email protected] Abstract: Foreign governments have used social media to influence politics in a range of countries by promoting propaganda, advocating controversial viewpoints, and spreading disinformation. We analyze 53 distinct foreign influence efforts (FIEs) targeting 24 different countries from 2013 through 2018. FIEs are defined as (i) coordinated campaigns by one state to impact one or more specific aspects of politics in another state (ii) through media channels, including social media, (iii) by producing content designed to appear indigenous to the target state. The objective of such campaigns can be quite broad and to date have included influencing political decisions by shaping election outcomes at various levels, shifting the political agenda on topics ranging from health to security, and encouraging political polarization. We draw on more than 460 media reports to identify FIEs, track their progress, and classify their features. Introduction Information and Communications Technologies (ICTs) have changed the way people communi- cate about politics and access information on a wide range of topics (Foley 2004, Chigona et al. 2009). Social media in particular has transformed communication between leaders and voters by enabling direct politician-to-voter engagement outside traditional avenues, such as speeches and press conferences (Ott 2017). In the 2016 U.S. presidential election, for example, social media platforms were more widely viewed than traditional editorial media and were central to the campaigns of both Democratic candidate Hillary Clinton and Republican candidate Donald Trump (Enli 2017).
    [Show full text]
  • Empirical Comparative Analysis of Bot Evidence in Social Networks
    Bots in Nets: Empirical Comparative Analysis of Bot Evidence in Social Networks Ross Schuchard1(&) , Andrew Crooks1,2 , Anthony Stefanidis2,3 , and Arie Croitoru2 1 Computational Social Science Program, Department of Computational and Data Sciences, George Mason University, Fairfax, VA 22030, USA [email protected] 2 Department of Geography and Geoinformation Science, George Mason University, Fairfax, VA 22030, USA 3 Criminal Investigations and Network Analysis Center, George Mason University, Fairfax, VA 22030, USA Abstract. The emergence of social bots within online social networks (OSNs) to diffuse information at scale has given rise to many efforts to detect them. While methodologies employed to detect the evolving sophistication of bots continue to improve, much work can be done to characterize the impact of bots on communication networks. In this study, we present a framework to describe the pervasiveness and relative importance of participants recognized as bots in various OSN conversations. Specifically, we harvested over 30 million tweets from three major global events in 2016 (the U.S. Presidential Election, the Ukrainian Conflict and Turkish Political Censorship) and compared the con- versational patterns of bots and humans within each event. We further examined the social network structure of each conversation to determine if bots exhibited any particular network influence, while also determining bot participation in key emergent network communities. The results showed that although participants recognized as social bots comprised only 0.28% of all OSN users in this study, they accounted for a significantly large portion of prominent centrality rankings across the three conversations. This includes the identification of individual bots as top-10 influencer nodes out of a total corpus consisting of more than 2.8 million nodes.
    [Show full text]
  • Hacks, Leaks and Disruptions | Russian Cyber Strategies
    CHAILLOT PAPER Nº 148 — October 2018 Hacks, leaks and disruptions Russian cyber strategies EDITED BY Nicu Popescu and Stanislav Secrieru WITH CONTRIBUTIONS FROM Siim Alatalu, Irina Borogan, Elena Chernenko, Sven Herpig, Oscar Jonsson, Xymena Kurowska, Jarno Limnell, Patryk Pawlak, Piret Pernik, Thomas Reinhold, Anatoly Reshetnikov, Andrei Soldatov and Jean-Baptiste Jeangène Vilmer Chaillot Papers HACKS, LEAKS AND DISRUPTIONS RUSSIAN CYBER STRATEGIES Edited by Nicu Popescu and Stanislav Secrieru CHAILLOT PAPERS October 2018 148 Disclaimer The views expressed in this Chaillot Paper are solely those of the authors and do not necessarily reflect the views of the Institute or of the European Union. European Union Institute for Security Studies Paris Director: Gustav Lindstrom © EU Institute for Security Studies, 2018. Reproduction is authorised, provided prior permission is sought from the Institute and the source is acknowledged, save where otherwise stated. Contents Executive summary 5 Introduction: Russia’s cyber prowess – where, how and what for? 9 Nicu Popescu and Stanislav Secrieru Russia’s cyber posture Russia’s approach to cyber: the best defence is a good offence 15 1 Andrei Soldatov and Irina Borogan Russia’s trolling complex at home and abroad 25 2 Xymena Kurowska and Anatoly Reshetnikov Spotting the bear: credible attribution and Russian 3 operations in cyberspace 33 Sven Herpig and Thomas Reinhold Russia’s cyber diplomacy 43 4 Elena Chernenko Case studies of Russian cyberattacks The early days of cyberattacks: 5 the cases of Estonia,
    [Show full text]
  • Artificial Intelligence and Countering Violent Extremism: a Primer
    Artificial Intelligence and Countering Violent Extremism: A Primer ByMarie Hxxnry Schroeter Mxxhm, Inxxs Oxxlxxmxxnxx, xxnd Fxxnxx Sxxngxxr GNET xxsis a xxspecial spxxcxxl project prxxjxxct delivered dxxlxxvxxrxxd by the International by thxx Intxxrnxxtxxnxxl Centre Cxxntrxx fxxrfor the thxx Study Stxxdy of Radicalisation,xxf Rxxdxxcxxlxxsxxtxxn, King’s College Kxxng’s London. Cxxllxxgxx Lxxndxxn. The author of this report is Marie Schroeter, Mercator Fellow on New Technology in International Relations: Potentials and Limitations of Artifical Intelligence to Prevent Violent Extremism Online The Global Network on Extremism and Technology (GNET) is an academic research initiative backed by the Global Internet Forum to Counter Terrorism (GIFCT), an independent but industry‑funded initiative for better understanding, and counteracting, terrorist use of technology. GNET is convened and led by the International Centre for the Study of Radicalisation (ICSR), an academic research centre based within the Department of War Studies at King’s College London. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing those, either expressed or implied, of GIFCT, GNET or ICSR. We would like to thank Tech Against Terrorism for their support with this report. CONTACT DETAILS For questions, queries and additional copies of this report, please contact: ICSR King’s College London Strand London WC2R 2LS United Kingdom T. +44 20 7848 2098 E. [email protected] Twitter: @GNET_research Like all other GNET publications, this report can be downloaded free of charge from the GNET website at www.gnet‑research.org. © GNET Artificial Intelligence and Countering Violent Extremism: A Primer Executive Summary Radicalisation can take place offline as well as online.
    [Show full text]
  • Developing a Twitter Bot That Can Join a Discussion Using State-Of-The-Art Architectures
    DEVELOPING A TWITTER BOT THAT CAN JOIN A DISCUSSION USING STATE-OF-THE-ART ARCHITECTURES A THESIS SUBMITTED TO THE GRADUATE SCHOOL OF NATURAL AND APPLIED SCIENCES OF MIDDLE EAST TECHNICAL UNIVERSITY BY YUSUF MÜCAHIT˙ ÇETINKAYA˙ IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF MASTER OF SCIENCE IN COMPUTER ENGINEERING AUGUST 2019 Approval of the thesis: DEVELOPING A TWITTER BOT THAT CAN JOIN A DISCUSSION USING STATE-OF-THE-ART ARCHITECTURES submitted by YUSUF MÜCAHIT˙ ÇETINKAYA˙ in partial fulfillment of the require- ments for the degree of Master of Science in Computer Engineering Department, Middle East Technical University by, Prof. Dr. Halil Kalıpçılar Dean, Graduate School of Natural and Applied Sciences Prof. Dr. Mehmet Halit Oguztüzün˘ Head of Department, Computer Engineering Prof. Dr. Ismail˙ Hakkı Toroslu Supervisor, Computer Engineering, METU Examining Committee Members: Prof. Dr. Tolga Can Computer Engineering, METU Prof. Dr. Ismail˙ Hakkı Toroslu Computer Engineering, METU Prof. Dr. Ahmet Co¸sar Computer Engineering, UTAA Date: I hereby declare that all information in this document has been obtained and presented in accordance with academic rules and ethical conduct. I also declare that, as required by these rules and conduct, I have fully cited and referenced all material and results that are not original to this work. Name, Surname: Yusuf Mücahit Çetinkaya Signature : iv ABSTRACT DEVELOPING A TWITTER BOT THAT CAN JOIN A DISCUSSION USING STATE-OF-THE-ART ARCHITECTURES Çetinkaya, Yusuf Mücahit M.S., Department of Computer Engineering Supervisor: Prof. Dr. Ismail˙ Hakkı Toroslu August 2019, 118 pages Twitter is today mostly used for sharing and commenting about news [1].
    [Show full text]
  • WEAPONS of MASS DISTRACTION: Foreign State-Sponsored Disinformation in the Digital Age
    WEAPONS OF MASS DISTRACTION: Foreign State-Sponsored Disinformation in the Digital Age MARCH 2019 PARK ADVISORS | Weapons of Mass Distraction: Foreign State-Sponsored Disinformation in the Digital Age Authored by Christina Nemr and William Gangware Acknowledgements The authors are grateful to the following subject matter experts who provided input on early drafts of select excerpts: Dr. Drew Conway, Dr. Arie Kruglanski, Sean Murphy, Dr. Alina Polyakova, and Katerina Sedova. The authors also appreciate the contributions to this paper by Andrew Rothgaber and Brendan O’Donoghue of Park Advisors, as well as the editorial assistance provided by Rhonda Shore and Ryan Jacobs. This report was produced with support from the US Department of State’s Global Engagement Center. Any views expressed in this report are those of the authors and do not necessarily reflect the views of the US State Department, Park Advisors, or its subject matter expert consultants. Any errors contained in this report are the authors’ alone. PARK ADVISORS | Weapons of Mass Distraction: Foreign State-Sponsored Disinformation in the Digital Age 0. Table of Contents 01 Introduction and contextual analysis 04 How do we define disinformation? 06 What psychological factors drive vulnerabilities to disinformation and propaganda? 14 A look at foreign state-sponsored disinformation and propaganda 26 Platform-specific challenges and efforts to counter disinformation 39 Knowledge gaps and future technology challenges PARK ADVISORS | Weapons of Mass Distraction: Foreign State-Sponsored Disinformation in the Digital Age 1 Introduction and 1. contextual analysis On July 12, 2014, viewers of Russia’s main state-run television station, Channel One, were shown a horrific story.
    [Show full text]
  • 2021-CICS-Horizon-Report Master.Pdf
    LEADERSHIP TEAM Dr. Dennis A. Trinkle, Director, Center for Information and Communication Sciences Dennis Trinkle is the Director for the Center of Information and Communication Sciences (CICS) and Director of the Applied Research Institutes at Ball State University. Dr. Trinkle brings a diverse multi-sector leadership background to his leadership of CICS, where he also holds the rank of Professor of Information Sciences and Communication. Prior to joining Ball State, Dr. Trinkle served as the system chief executive, provost, and chief academic officer for Harrison College. Dr. Trinkle has served leading institutions across the higher education, corporate, and government sectors, including serving as the chief executive officer of IHETS, chief information officer, associate vice president for academic affairs, and Tenzer Professor of Infor- mation Technology for DePauw University, and president, chief executive officer, and board chair for many companies and non-profit organizations. He is the author and editor of 17 books and more than 50 articles on entrepreneurship, technology, leadership, teaching and learning, and history. Together with Kevin Allen, CEO of EiGames, he developed the award-winning video game on the entrepreneurial process—Fireworks: The Entrepreneurship Experience (2018). His latest book: Recipes for Success: How to Start a Business was published in spring 2021. Christopher Nouhan, Graduate Student, Center for Information and Communication Sciences Christopher Nouhan is a Graduate Assistant for the Center for Information and Communication Sciences (CICS) at Ball State University. Prior to enrolling in CICS, Christopher attended Alma College, where he obtained a B.A. in New Media Studies. While at Alma College, he founded Bitworks Productions, a pre-pro- fessional program that bridged media students with community clients, providing professional development opportunities to his peers.
    [Show full text]
  • World Changing Ideas for 2018
    INSIDE NIKE’S DIGITAL PLAYBOOK / NINTENDO SWITCHES IT UP, AGAIN WORLD CHANGING IDEAS TWITTER FOR 2018 L’Oréal, The North Face, FACES L.A. Mayor Eric Garcetti, THE and more TRUTH The Creative HOW JACK DORSEY’S Minds Behind FREE-SPEECH PLATFORM Casper,Allbirds WAS HIJACKED & Birchbox BY THE DARK SIDE BY AUSTIN CARR AND HARRY McCRACKEN We believe that there is no mold for invention. The proof? The Genesis G80 Sport. With its impeccably crafted balance of power and detail, aggressive dark-chrome crosshatched front grille, carbon-fiber-trimmed dash, and 19-inch copper accented alloy wheels artfully surrounding a 365-hp twin-turbo engine, it’s sure to pin you to its copper stitched sport seats. Add in the peace of mind of our most advanced safety features standard and Complimentary Service Valet, and the only thing the G80 Sport doesn’t have is limits. GENESIS.COM Optional features shown. 365-hp rating with premium fuel only. Service Valet available from Authorized Genesis Retailers during complimentary maintenance period. Includes normal wear, parts replacement and warranty repairs for 3 years or 36,000 miles, whichever comes first. Coverage area varies by retailer. Charges and fees may apply. Applies to original owner/lessee only. AS A RULE, CHALLENGE THE RULE. WON’T PAY FOR ITSELF. Switch to GEICO and save money for the things you love. Maybe it’s that high-end client dinner. Or the last-minute flight overseas. Doing business is what you love – and it doesn’t come cheap. So switch to GEICO, because you could save 15% or more on car insurance.
    [Show full text]
  • Trends in Online Influence Efforts
    Trends in Online Influence Efforts∗ Diego A. Martiny Jacob N. Shapiroz Julia G. Ilhardtx Version 2.0 August 5, 2020 Abstract Information and Communications Technologies (ICTs) create novel opportuni- ties for a wide range of political actors. Foreign governments have used social media to influence politics in a range of countries by promoting propaganda, advocating controversial viewpoints, and spreading disinformation. This report updates previ- ous work with data on 76 such foreign influence efforts (FIE) targeting 30 different countries from 2013 through 2019, as well as 20 domestic influence efforts (DIE) in which governments targeted their own citizens. Influence efforts (IEs) are de- fined as: (i) coordinated campaigns by a state or the ruling party in an autocracy to impact one or more specific aspects of politics at home or in another state, (ii) through media channels, including social media, by (iii) producing content designed to appear indigenous to the target state. The objective of such campaigns can be quite broad and to date have included shaping election outcomes at various levels, shifting the political agenda on topics ranging from health to security, and encour- aging political polarization. Our data draw on more than 920 media reports and 380 research articles/reports to identify IEs, track their progress, and classify their features. ∗We are grateful to a range of colleagues including Laura Courchesne, Nick Feamster, Andy Guess, Hans Klein, Brendan Stewart, and Alicia Wanless for helpful comments and feedback. Will Lowe provided invaluable advice on data structure and data entry. Jordan Allen, Nicola Bariletto, Arya Goel, Danielle Hull, Janette Lu, Imane Mabrouk, Matthew Merrigan, Justinas Mickus, Brendan O'Hara, Jan Oledan, Kamila Radjabova, Nina Sheridan, Joe Shipley, Jack Tait, Kamya Yadav, and Luca Zanotti provided excellent research assistance.
    [Show full text]
  • "Deep Faking" Political Twitter Using Transfer Learning and GPT-2
    "Deep Faking" Political Twitter using Transfer learning and GPT-2 Ryan Ressmeyer Sam Masling Department of Electrical Engineering Department of Computer Science Stanford University Stanford University [email protected] [email protected] Madeline Liao Department of Computer Science Stanford University [email protected] Abstract In the modern political climate, Twitter has become one of, if not the most impactful mediums for both current and aspiring politicians to communicate with their constituents. Our group attempted to replicate a variety of politicians’ twitter accounts using the recently released GPT-2 model from OpenAI and transfer learning. We compared our results to former state-of-the-art LSTM recurrent models to illustrate the advantages of transfer learning. Additionally, we used GPT-2 to extrapolate from a given context, allowing for robust tweet mapping from one user to another. 1 Introduction In recent election years, Twitter has become an increasingly important communication platform for American politicians. Our goal was to see if we could use state-of-the-art NLP techniques to not only recognize the different style and contents of tweets written by different politicians, but to actually generate similar tweets. This idea stemmed from two popular fields of study within NLP: Twitter sentiment analysis, and zero-shot text generation. By utilizing transfer learning and building upon OpenAI’s GPT-2 text generation model, we sought to train a model for the domain-specific task of generating tweets based on a given account. OpenAI’s GPT-2 is a transformer-based language model trained with the goal of predicting the next word given all previous words of some text [8].
    [Show full text]
  • A Comparative Study of Bot Detection Techniques Methods with an Application Related to Covid-19 Discourse on Twitter
    A comparative study of Bot Detection techniques methods with an application related to Covid-19 discourse on Twitter. Marzia Antenore∗1, Jos´eM. Camacho-Rodr´ıguez †2, Emanuele Panizzi‡2 1Department of Communication and Social Research, Sapienza University of Rome 2Department of Computer Science, Sapienza University of Rome Abstract Bot Detection is an essential asset in a period where Online Social Networks(OSN) is a part of our lives. This task becomes more relevant in crises, as the Covid-19 pandemic, where there is an incipient risk of proliferation of social bots, producing a possible source of misinformation. In order to address this issue, it has been compared different methods to detect automatically social bots on Twitter using Data Selection. The techniques utilized to elaborate the bot detection models include the utilization of features as the tweets' metadata or the Digital Fingerprint of the Twitter accounts. In addition, it was analyzed the presence of bots in tweets from different periods of the first months of the Covid-19 pandemic, using the bot detection technique which best fits the scope of the task. Moreover, this work includes also analysis over aspects regarding the discourse of bots and humans, such as sentiment or hashtag utilization. arXiv:2102.01148v1 [cs.SI] 1 Feb 2021 1 Introduction According to [47], a bot is a socio-technical entity based on a software program whose aim is to simulate human behavior in Online Social Networks(OSN) such as Facebook, Twitter, or Instagram. Bots are configured to resemblance as humans not just to other human users, but also to the OSN platform [28].
    [Show full text]