Download Download

Download Download

Proceedings of the Fifteenth International AAAI Conference on Web and Social Media (ICWSM 2021) Uncovering Coordinated Networks on Social Media: Methods and Case Studies Diogo Pacheco,∗1,2 Pik-Mai Hui,∗1 Christopher Torres-Lugo,∗1 Bao Tran Truong,1 Alessandro Flammini,1 Filippo Menczer1 1Observatory on Social Media, Indiana University Bloomington, USA 2Department of Computer Science, University of Exeter, UK [email protected],fhuip,torresch,baotruon,aflammin,fi[email protected] Abstract develop software to impersonate users and hide the iden- tity of those who control these social bots — whether they Coordinated campaigns are used to influence and manipulate are fraudsters pushing spam, political operatives amplifying social media platforms and their users, a critical challenge to misleading narratives, or nation-states waging online war- the free exchange of information online. Here we introduce a general, unsupervised network-based methodology to un- fare (Ferrara et al. 2016). Cognitive and social biases make cover groups of accounts that are likely coordinated. The pro- us even more vulnerable to manipulation by social bots: posed method constructs coordination networks based on ar- limited attention facilitates the spread of unchecked claims, bitrary behavioral traces shared among accounts. We present confirmation bias makes us disregard facts, group-think five case studies of influence campaigns, four of which in the and echo chambers distort perceptions of norms, and the diverse contexts of U.S. elections, Hong Kong protests, the bandwagon effect makes us pay attention to bot-amplified Syrian civil war, and cryptocurrency manipulation. In each memes (Weng et al. 2012; Hills 2019; Ciampaglia et al. of these cases, we detect networks of coordinated Twitter 2018; Lazer et al. 2018; Pennycook et al. 2019). accounts by examining their identities, images, hashtag se- Despite advances in countermeasures such as machine quences, retweets, or temporal patterns. The proposed ap- proach proves to be broadly applicable to uncover different learning algorithms and human fact-checkers employed by kinds of coordination across information warfare scenarios. social media platforms to detect misinformation and inau- thentic accounts, malicious actors continue to effectively de- ceive the public, amplify misinformation, and drive polar- Introduction ization (Barrett 2019). We observe an arms race in which Online social media have revolutionized how people access the sophistication of attacks evolves to evade detection. news and information, and form opinions. By enabling ex- Most machine learning tools to combat online abuse tar- changes that are unhindered by geographical barriers, and get the detection of social bots, and mainly use meth- by lowering the cost of information production and con- ods that focus on individual accounts (Davis et al. 2016; sumption, social media have enormously broadened partici- Varol et al. 2017; Yang et al. 2019; 2020; Sayyadiharikandeh pation in civil and political discourse. Although this could et al. 2020). However, malicious groups may employ coor- potentially strengthen democratic processes, there is in- dination tactics that appear innocuous at the individual level, creasing evidence of malicious actors polluting the informa- and whose suspicious behaviors can be detected only when tion ecosystem with disinformation and manipulation cam- observing networks of interactions among accounts. For in- paigns (Lazer et al. 2018; Vosoughi, Roy, and Aral 2018; stance, an account changing its handle might be normal, but Bessi and Ferrara 2016; Shao et al. 2018; Ferrara 2017; a group of accounts switching their names in rotation is un- Stella, Ferrara, and De Domenico 2018; Deb et al. 2019; likely to be coincidental. Bovet and Makse 2019; Grinberg et al. 2019). Here we propose an approach to reveal coordinated be- While influence campaigns, misinformation, and propa- haviors among multiple actors, regardless of their auto- ganda have always existed (Jowett and O’Donnell 2018), mated/organic nature or malicious/benign intent. The idea social media have created new vulnerabilities and abuse op- is to extract features from social media data to build a co- portunities. Just as easily as like-minded users can connect ordination network, where two accounts have a strong tie if in support of legitimate causes, so can groups with fringe, they display unexpectedly similar behaviors. These similar- conspiratorial, or extremist beliefs reach critical mass and ities can stem from any metadata, such as content entities become impervious to expert or moderating views. Platform and profile features. Networks provide an efficient represen- APIs and commoditized fake accounts make it simple to tation for sparse similarity matrices, and a natural way to detect significant clusters of coordinated accounts. Our main ∗Equal contributions. contributions are: Copyright c 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. • We present a general approach to detect coordination, 455 which can in principle be applied to any social media plat- to identify coordinated accounts during the 2012 election in form where data is available. Since the method is com- South Korea (Keller et al. 2017; 2019). pletely unsupervised, no labeled training data is required. While these approaches can work well, each is designed • Using Twitter data, we present five case studies by in- to consider only one of the many possible coordination di- stantiating the approach to detect different types of co- mensions. Furthermore, they are focused on coordination ordination based on (i) handle changes, (ii) image shar- features that are likely observed among automated accounts; ing, (iii) sequential use of hashtags, (iv) co-retweets, and inauthentic coordination among human-controlled accounts (v) synchronization. is also an important challenge. The unsupervised approach proposed here is more general in allowing multiple similar- • The case studies illustrate the generality and effectiveness ity criteria that can detect human coordination in addition to of our approach: we are able to detect coordinated cam- bots. As we will show, several of the aforementioned unsu- paigns based on what is presented as identity, shown in pervised methods can be considered as special cases of the pictures, written in text, retweeted, or when these actions methodology proposed here. are taken. • We show that coordinated behavior does not necessarily Methods imply automation. In the case studies, we detected a mix The proposed approach to detect accounts acting in coor- of likely bot and human accounts working together in ma- dination on social media is illustrated in Fig. 1. It can be licious campaigns. described by four phases: • Code and data are available at github.com/IUNetSci/ 1. Behavioral trace extraction: The starting point of coor- coordination-detection to reproduce the present results dination detection should be a conjecture about suspicious and apply our methodology to other cases. behavior. Assuming that authentic users are somewhat in- dependent of each other, we consider a surprising lack Related Work of independence as evidence of coordination. The imple- Inauthentic coordination on social media can occur among mentation of the approach is guided by a choice of traces social bots as well as human-controlled accounts. How- that capture such suspicious behavior. For example, if we ever, most research to date has focused on detecting social conjecture that accounts are controlled by an entity with bots (Ferrara et al. 2016). Supervised machine learning mod- the goal of amplifying the exposure of a disinformation els require labeled data describing how both humans and source, we could extract shared URLs as traces. Coordi- bots behave. Researchers created datasets using automated nation scenarios may be associated with a few broad cat- honeypot methods (Lee, Eoff, and Caverlee 2011), human egories of suspicious traces: annotation (Varol et al. 2017), or likely botnets (Echeverria, (a) Content: if the coordination is based on the content Besel, and Zhou 2017; Echeverria and Zhou 2017). These being shared, suspicious traces may include words, n- datasets have proven useful in training supervised mod- grams, hashtags, media, links, user mentions, etc. els for bot detection (Davis et al. 2016; Varol et al. 2017; (b) Activity: coordination could be revealed by spatio- Yang et al. 2019). temporal patterns of activity. Examples of traces that One downside of supervised detection methods is that can reveal suspicious behaviors are timestamps, places, by relying on features from a single account or tweet, and geo-coordinates. they are not as effective at detecting coordinated accounts. (c) Identity: accounts could coordinate on the basis of per- This limitation has been explored in the context of detect- sonas or groups. Traces of identity descriptors could ing coordinated social bots (Chen and Subramanian 2018; be used to detect these kinds of coordination: name, Cresci et al. 2017; Grimme, Assenmacher, and Adam 2018). handle, description, profile picture, homepage, account The detection of coordinated accounts requires a shift to- creation date, etc. ward the unsupervised learning paradigm. Initial applica- tions focused on clustering or community detection algo- (d) Combination: the detection of coordination might re- combination rithms in an attempt to identify similar features among quire a of multiple dimensions. For in- pairs of accounts

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    12 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us