Bias Misperceived: the Role of Partisanship and Misinformation in Youtube Comment Moderation

Bias Misperceived: the Role of Partisanship and Misinformation in Youtube Comment Moderation

Proceedings of the Thirteenth International AAAI Conference on Web and Social Media (ICWSM 2019) Bias Misperceived: The Role of Partisanship and Misinformation in YouTube Comment Moderation Shan Jiang, Ronald E. Robertson, Christo Wilson Northeastern University, USA fsjiang, rer, [email protected], Abstract censoring their views (Kamisar 2018; Usher 2018). Two US House Committees have held hearings on content mod- Social media platforms have been the subject of controversy eration practices to “specifically look at concerns regard- and scrutiny due to the spread of hateful content. To address ing a lack of transparency and potential bias in the filter- this problem, the platforms implement content moderation us- ing a mix of human and algorithmic processes. However, con- ing practices of social media companies (Facebook, Twit- tent moderation itself has lead to further accusations against ter and YouTube)” (Bickert, Downs, and Pickles 2018; the platforms of political bias. In this study, we investigate Dorsey 2018). These concerns are driven by multiple fac- how channel partisanship and video misinformation affect tors, including anecdotal reports that: Facebook’s Trending the likelihood of comment moderation on YouTube. Using News team did not promote stories from conservative media a dataset of 84,068 comments on 258 videos, we find that outlets (Nunez 2016), Twitter “shadow banned” conserva- although comments on right-leaning videos are more heav- tive users (Newton 2018), fact-checking organizations are ily moderated from a correlational perspective, we find no biased (Richardson 2018), and selective reporting by partisan evidence to support claims of political bias when using a news agencies (Arceneaux, Johnson, and Murphy 2012). causal model that controls for common confounders (e.g., hate However, there is, to our knowledge, no scientific evidence speech). Additionally, we find that comments are more likely to be moderated if the video channel is ideologically extreme, that social media platforms’ content moderation practices if the video content is false, and if the comments were posted exhibit systematic partisan bias. On the contrary, there are after a fact-check. many cases where ideologically liberal users were moder- ated, although these cases have received less attention in the media (Masnick 2018). It is possible that moderation only 1 Introduction appears to be biased because political valence is correlated In the wake of 2016–2017 global election cycle, social me- with other factors that trigger moderation, such as bullying, dia platforms have been subject to heightened levels of con- calls to violence, or hate speech (Gillespie 2018). Further, troversy and scrutiny. Besides well-documented privacy is- there is evidence suggesting that users tend to overestimate sues (Solon 2018), social media platforms have been increas- bias in moderation decisions (Shen et al. 2018). ingly used to promote partisanship (Allcott and Gentzkow In this study, we use YouTube as a lens and take a first 2017), spread misinformation (Constine 2018), and breed step towards disentangling these issues by investigating how violent hate speech (Olteanu et al. 2018). partisanship and misinformation in videos affect the like- The solution promulgated by social media platforms for lihood of comment moderation. Specifically, we examine these ills is an increase in content moderation. In terms of four hypotheses related to four variables of YouTube videos: mechanisms, the major platforms have committed to hiring the direction of partisanship (H1a0: left/right), the magni- tens of thousands of new human moderators (Levin 2017), in- tude of partisanship (H1b0: extreme/center), the veracity vesting in more artificial intelligence to filter content (Gibbs of the content (H2a0: true/false), and whether a comment 2017), and partnering with fact-checking organizations to was posted before or after the video was fact-checked (H2b0: identify misinformation (Glaser 2018). In terms of policy, the before/after). For each variable, we start with the null hy- platforms are updating their community guidelines with ex- potheses (H0) that the variable has no effect on comment panded definitions of hate speech, harassment, etc (YouTube moderation, and then use correlational and casual models to 2018; Facebook 2018; Twitter 2018). collect evidence on rejecting the null hypotheses. The con- This increased reliance on content moderation faces a ceptual framework of our hypotheses is shown in Figure 1. backlash from ideological conservatives, who claim that To investigate these hypotheses, we collected a dataset social media platforms are biased against them and are of 84,068 comments posted across 258 YouTube videos,1 and associate them to partisanship labels from previous Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 1The dataset is available at: https://moderation.shanjiang.me 278 Author political ideology YouTube videos Fact-check articles Social engagement Partisan bias Misinformation User comments Left/Right Extreme/Center True/False Fact-checked/Not H1a 0 : H1b 0 : Human and algorithmic process H2a 0 : H2b 0 : not rejected rejected rejected rejected Content moderation Figure 1: Conceptual framework. We investigate the effect of partisanship (i.e., left/right, extreme/center) and misinformation (i.e., true/false, fact-checked/not) on comment moderation. Potential confounders include social engagement on YouTube videos (e.g., views and likes) and linguistics in comments (e.g., hate speech). research (Robertson et al. 2018) and misinformation la- on YouTube (e.g., a video or comment) is judged to violate bels from Snopes or PolitiFact (Jiang and Wilson 2018). the guidelines, it is taken down, i.e., moderated. We first run a correlational analysis and find that all of There are multiple reasons why a comment could be mod- our hypothesized variables significantly correlate with the erated on YouTube. A comment may be reviewed by pa- likelihood of comment moderation, suggesting a politi- trolling YouTube moderators, or a comment may be flagged cal bias against right-leaning content. However, we argue by YouTube users and then reviewed by the YouTube modera- that such bias is misperceived as it ignores other con- tors (Levin 2017). Additionally, a comment may be removed founding variables that potentially contribute to modera- by the corresponding video uploader, or by the commenter tion decisions, such as social engagement (e.g., views and themselves (YouTube 2018). Besides these human efforts, likes) (Mondal, Silva, and Benevenuto 2017) and the lin- YouTube also uses algorithms that automatically flag and guistics in comments (e.g., hate speech) (Shen et al. 2018; moderate inappropriate content (Gibbs 2017). In general, the Chandrasekharan et al. 2018). Therefore, we re-analyze our mechanisms that lead to comment moderation are convo- dataset using a causal propensity score model to investigate luted. Therefore, we view the internal YouTube system as a null hypotheses when potential confounds are controlled. We black-box, and focus on the outcome of moderation instead. make the following observations: • H1a0: not rejected. No significant difference is found for Pros & Cons of Content Moderation comment moderation on left- and right-leaning videos. Content moderation has been shown to have positive effects • H1b0: rejected. Comments on videos from ideologically on social media platforms. A study that investigated Red- extreme channels are ∼50% more likely to be moderated dit’s ban of the r/fatpeoplehate and r/CoonTown communities than center channels. found that the ban expelled more “bad actors” than expected, • H2a0: rejected. Comments on true videos are ∼60% less and those who stayed posted much less hate speech than likely to be moderated than those on false videos. before the ban (Chandrasekharan et al. 2017). A study that • H2b0: rejected. Comments posted after a video is fact- interviewed users of Twitter’s “blocklist” feature discussed checked are ∼20% more likely to be moderated than those how it can be used to prevent harassment (Jhaver et al. 2018). posted before the fact-check. However, content moderation systems have also raised concerns about bias and efficacy. Human moderators have We approach these hypotheses using an empirical method been shown to bring their own biases into the content evalua- for auditing black-box decision-making processes (Sandvig tion process (Diakopoulos and Naaman 2011) and automated et al. 2014) based on publicly available data on YouTube. moderation algorithms are prone to false positives and nega- Neither we, nor the critics, have access to YouTube’s internal tives (Veletsianos et al. 2018). These moderation strategies systems, data, or deliberations that underpin moderation deci- are also brittle: a study on Instagram found that users in pro- sions. Instead, we aim to highlight the difference in perceived eating disorder communities invented variations of banned bias when analyzing available data using correlational and tags (e.g., “anorexie” instead of “anorexia”) to circumvent causal models, and further, foster a healthier discussion of lexicon-based moderation (Chancellor et al. 2016). algorithmic and human bias in social media. Researchers have also studied the community norms be- hind moderation from a linguistic perspective. A study on 2 Background & Related Work Reddit used 2.8M removed comments to identify macro-, Social media platforms publish sets of community

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    12 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us