Characterizing (Un)Moderated Textual Data in Social Systems

Characterizing (Un)Moderated Textual Data in Social Systems

Characterizing (Un)moderated Textual Data in Social Systems Lucas Lima∗, Julio C. S. Reis∗y, Philipe Melo∗, Fabr´ıcio Murai∗, Fabr´ıcio Benevenuto∗ ∗Universidade Federal de Minas Gerais (UFMG), Brazil, yUniversidade FUMEC, Brazil flucaslima, julio.reis, philipe, murai, [email protected] Abstract—Despite the valuable social interactions that online not allow illegal activity, spam, or form of illegal pornography, media promote, these systems provide space for speech that promotion of violence and terrorism. would be potentially detrimental to different groups of people. Despite existing recent efforts that attempt to understand The moderation of content imposed by many social media has motivated the emergence of a new social system for free speech the content shared in Gab [4, 5], there is still a need for an named Gab, which lacks moderation of content. This article char- understanding regarding the amount and forms of hate speech acterizes and compares moderated textual data from Twitter with in an unmoderated system such as Gab. In this article, we a set of unmoderated data from Gab. In particular, we analyze provide a diagnostic of hate in the unmoderated content from distinguishing characteristics of moderated and unmoderated Gab, by categorizing the different forms of hate speech in that content in terms of linguistic features, evaluate hate speech and its different forms in both environments. Our work shows that system and comparing it with Twitter, a proxy for a moderated unmoderated content presents different psycholinguistic features, system. Specifically, we identify textual characteristics of a more negative sentiment and higher toxicity. Our findings support set of unmoderated (or barely moderated) data, in this work that unmoderated environments may have proportionally more represented by Gab, and compare them with characteristics online hate speech. We hope our analysis and findings contribute of moderated data, here represented by Twitter. Our study is to the debate about hate speech and benefit systems aiming at deploying hate speech detection approaches. based on the analysis of 7; 794; 990 Gab posts and 9; 118; 006 Index Terms—Social Network, Moderated Content, Unmoder- tweets from August 2016 to August 2017. At a high level, our ated Content, Hate Speech, Gab, Twitter. analysis is centered around the following research questions: Research Question 1: What are the distinguishing characteris- tics of moderated content in Twitter and unmoderated content I. INTRODUCTION in Gab in terms of linguistic features, sentiment, and toxicity? The Web has changed the way our society communicates, Research Question 2: What are the most common types of giving rise to social platforms where users can share different hate in an unmoderated and moderated environment? types of content and freely express themselves through posts To answer our first research question we quantify the lin- containing personal opinions. Unfortunately, with the popular- guistic differences across moderated and unmoderated social ization of this new flavor of communication, toxic behaviors systems by applying established language processing tech- enacted by some users have been gaining prominence through niques. Our analysis shows that content in Gab and Twitter online harassment and hate speech. These platforms have have different linguistics patterns, with higher toxicity and become the stage for numerous cases of online hate speech, a a more negative overall sentiment score in the unmoderated type of discourse that aims at attacking a person or a group Gab content. Additionally, we find that, in general, Gab has on the basis of race, religion, ethnic origin, sexual orientation, more hate posts than Twitter. We show that Gender and Class disability, or gender [1]. types of hate are more frequent on Twitter, whereas Disability, Recently, to prevent the proliferation of toxic content, most Ethnicity, Sexual Orientation, Religion, and Nationality types arXiv:2101.00963v1 [cs.SI] 4 Jan 2021 online social networks prohibited hate speech in their user tend to appear proportionality more in Gab. policies and enforced this rule by deleting posts and banning These findings highlight the importance of creating moder- users who violate it. Particularly, Twitter, Facebook, and ation policies as an effort to fight online hate speech in social Google (YouTube) have largely increased removals of hate systems, and also point out possible points for improvement speech content [2] by making available their hate policies in the design of content policies for social media systems. so users can actively report content that might violate their Additionally, our findings suggest that the unmoderated con- policies. Reddit also deleted some communities related to fat- tent found in Gab might be an appropriate data source for shaming and hate against immigrants [3]. This scenario has the development of learning approaches to detect hate speech. motivated the emergence of a new social network system, Thus, as a final contribution, we make our hate-labeled Gab called Gab. In essence, Gab is very similar to Twitter, but posts available for the research community1 can foster the barely moderates any of the content shared by its users. development of future hate speech detection systems. According to Gab guidelines, the website promotes freedom of expression and states that “the only valid form of censorship is an individual’s own choice to opt-out”. They, however, do 1https://github.com/lhenriquecl/unmoderated-hate-dataset II. RELATED WORK III. DATASETS AND METHODS A. Social Media Content Moderation A. Datasets The ongoing discussion on content moderation in an online Our Gab dataset comprises posts from users crawled fol- environment led Internet companies to reflect about the unde- lowing a Breadth-First Search (BFS) scheme on the graph of sirable attention their sites can attract and the consequences followers and friends. We used as seeds users who authored 2 of it , thus the first steps towards regulating and moderating posts listed by categories on the Gab main page. We imple- content have already been taken. There is a lot of debate mented a distributed and parallel crawler that ran in August on how social media platforms moderate their own content, 2017. In total, our Gab dataset comprises 12; 829; 976 posts, and how their moderation policies are shaped. Twitter and obtained from 171; 920 users (the estimated number of users 3,4 YouTube, for instance, make available their hate policies in August 2017 was 225 thousand [15]). so users can actively report content that might violate their The Twitter dataset contains English posts randomly se- policies. Our work contributes to this discussion since we lected from the Twitter 1% Streaming API. For consolidating quantify the differences between moderated and unmoderated our dataset and keep data consistency, we consider only text data as an effort to shed light on the importance of creating random tweets published in the same period as Gab posts, different methodologies and policies to outline the boundaries which gives us also 12; 829; 976 tweets. After preprocessing of hate in social media. and removing duplicated posts in both datasets, we have a total Besides, some studies have focused on the understanding of 7,794,990 Gab posts and 9,118,006 tweets. These are the of systems that lack moderation of content. Finkelstein et final sets of posts for each media that are going to be further al. [6] focus on making an extensive analysis of antisemitism analyzed in this work. in 4chan’s and Gab. Their results provide a quantitative data- driven framework for understanding this form of offensive B. Language Processing Methods content. Zannettou et. al [5] and Lima et. al [4] present the first Next, we describe the methods used in this work to perform characterization studies on Gab, analyzing network structure, language characterization. users, and posts. Our effort is complementary to these works 1) Linguistic Analysis: One of our goals is to understand as we compare the textual data shared from an unmoderated the distinguishing linguistic characteristics of posts on Gab and system like Gab with a moderated one as Twitter and deeply Twitter and contrast them. Thus, we use the 2015 version of investigate the types of hate in both social systems. the Linguistic Inquiry and Word Count (LIWC) [16] to extract B. Online Manifestations of Hate Speech and analyze the distribution of psycholinguistic elements posts of both media. LIWC is a psycholinguistic lexicon system that A vast number of studies were conducted to provide a better categorizes words into psychologically meaningful groups all understanding of hate speech on the Internet. Silva et al. and of which form the set of LIWC attributes, and has been widely Mondal et al. [7, 8] provide a deeper understanding of the used for several different tasks [17, 18, 19, 20]. hateful messages exchanged in social networks, studying who are the most common targets of hate speech in these systems. 2) Sentiment Analysis: We perform sentiment analysis on Salminen [9] create a taxonomy and use it to investigate how Gab and Twitter posts as a complementary effort to character- different features and algorithms can influence the results of ize the differences between moderated and unmoderated ac- the hatefulness classification of text using several machine counts. We use an established opinion mining method to mea- learning methods. Chandrasekharan et al. [10] characterize two sure sentiment score on

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    5 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us