ISSUE BRIEF: HOW DISINFORMATION IMPACTS POLITICS AND PUBLICS Prepared by Dean Jackson, International Forum for Democratic Studies IN THIS BRIEF: • How disinformation is used and consumed • Proactive and reactive disinformation strategies in different country contexts • The scale of the disinformation crisis EXPANDING THE ANALYTICAL FRAME Disinformation—the use of half-truth and non-rational argument to manipulate public opinion in pursuit of political objectives—is a growing threat to the public sphere in countries around the world. The challenge posed by Russian disinformation has attracted significant attention in the United States and Europe; over time, observers have noted its role in “hybrid warfare,” its use to degrade public trust in media and state institutions, and its ability to amplify social division, resentment, and fear. But Moscow is merely the most prominent purveyor of disinformation, not its sole source. Political actors around the world, ranging in size from state agencies to individuals, have found ways to exploit the economics of digital advertising and the fast- paced nature of the modern information ecosystem for their political advantage. Growing appreciation of the problem’s scale invites a shift in frame: from national security threat from a discrete actor to a broader appreciation of political-economic weaknesses in the contemporary information space. Disinformation has a wider variety of purposes, in a wider variety of settings, than is commonly appreciated. In the short term, it can be used to distract from an issue, obscure the truth, or to inspire its consumers to take a certain course of action. In the long-term, disinformation can be part of a strategy to shape the information environment in which individuals, governments, and other actors form beliefs and make decisions. DISINFORMATION AS A REACTIVE TACTIC In the short term, disinformation can be utilized reactively by different entities: for example, when Russian-backed fighters in Eastern Ukraine shot down a commercial airliner, Russian state media went into overdrive proposing multiple, often conflicting alternative explanations for the plane’s crash. Disinformation’s applications have also been evident in Syria, where Russian diplomats, media, and intelligence services have falsified evidence, pushed misleading narratives, and spread falsehoods relating to the role of Russia’s airstrikes, as well as to obscure evidence of the Syrian government’s use of chemical weapons. Another common technique is to react to a crisis by flooding the information space and drowning out discussion. After opposition protests broke out in Syria during 2011, newly-created Twitter accounts began harassing Syrian users, and social media researches allege that the Assad regime paid a public relations firm to flood opposition hashtags with photos of nature scenery and sports scores. BOTS AND TROLLS SHAPE POLITICAL CONVERSATION ONLINE Online trolling, harassment, and distraction—especially by highly active automated accounts—are a key component of the modern disinformation purveyor’s toolkit. These techniques push independent voices out of public spaces and are sometimes considered a new form of political censorship. The Chinese Communist Party (CCP) was an early pioneer of this approach: for at least a decade, Beijing has deployed a “fifty-cent party” (apocryphally named for posters’ going rate per post) to “astroturf” support for the government and derail online political conversations that could spark mass mobilization. Recent estimates suggest this effort encompasses two million individuals, many of them state employees, and produces nearly 450 million social media posts per year. Over time, similar approaches became a common aspect of authoritarian information manipulation and were later amplified through automation. In the early to mid-2000s, the Russian government began recruiting human commenters before later adopting the use of automated “bot” accounts. One study suggests that on Twitter more than half of tweets in Russian are produced by automated accounts. Aiming to avoid detection, many disinformation campaigns now avail themselves of accounts that are partially automated, partially controlled by human users; these are often referred to as cyborg or sock puppet accounts. In recent years, the use of bots and trolls to shape online discussion has become so common across countries that it could be considered a widely exploited bug in the digital public square extending far beyond conflict or authoritarian settings. In Mexico, paid political consultants orchestrated the theft of campaign secrets and the large scale distribution of disinformation to voters. Such activity continues to this day, as pro-government accounts swarm political hashtags, threaten the lives of activists, and marginalize protesters. In the Philippines, where the public square faces significant threats both online and off, interview-based research has explored a sophisticated “underground” public relations industry in which digital strategists, social media influencers, and paid commenters compete to deliver their clients the greatest degree of control over political narratives on the internet. In a stroke of market innovation, the subcontracting of digital disinformation in the Philippines has tied the financial and career incentives of competing freelancers to the objectives of national political parties, to devastating effect. PROACTIVE DISINFORMATION AND THE “DEMAND SIDE” OF THE CHALLENGE The effectiveness of ‘reactive’ disinformation is limited by the unpredictability of real- world events. While it can offer those who use it a lifeline in times of crisis, reactive disinformation is by definition a response to unexpected, uncontrollable, or undesirable events and therefore generally used by those in disadvantageous strategic positions. Used proactively, disinformation provides much greater potential to move audiences to action, shape or confuse public understanding, and influence political events. However, it does not provide a blank canvas on which to work. Effective disinformation campaigns usually draw on preexisting divides within target societies and produce content for which there is societal demand. Disinformation is at its most dangerous when amplifying existing political beliefs and divisions as opposed to introducing new beliefs or narratives into the public sphere. It is effective in doing so in part due to low trust in media and in part due to cognitive biases that make many consumers more likely to believe content that confirms their beliefs, to prefer partisan cheerleading over the conclusions of fact-checkers, and to share content that makes them angry or afraid. Research into the impact of social media use on political polarization is ongoing, but at a minimum suggests that the emergence of social media platforms as news sources has diminished the power of traditional “gatekeepers” of news and information. In turn, social media seems to have increased the social and political influence of a voracious subset of news consumers engaged in “motivated reasoning”—the selected interpretation of information to justify one’s preexisting beliefs, stances, or desires. These factors, combined with the speed at which information spreads online, create ideal conditions for disinformation campaigns. DIGITAL DISINFORMATION CAN INSPIRE REAL-WORLD ACTION Proactive disinformation campaigns can achieve real-world impact by influencing the actions of its consumers. A prominent example comes from Germany’s 2016 “Lisa case,” which ignited nationwide debate over the country’s resettlement of Middle Eastern refugees and offered Moscow an opportunity to stoke divisions within Germany. Lisa, a thirteen year-old Russian-German girl, alleged that two migrant men kidnapped and raped her; the allegations were later proven to be untrue, but not before Russian state media actively spread the story and the Russian Foreign Minister publicly accused Berlin of a cover up. In Germany, thousands protested the government’s handling of the case. By using media and diplomatic resources to promote a false story at a time of rising German anti-migrant sentiment, Moscow sought to exploit domestic German political divides to encourage mass demonstrations and damage the German government politically. Digital disinformation often promotes xenophobic sentiment, and hate speech is common. In India, far-right religious figures used messaging applications to spread false claims about religious minorities, sparking communal violence. In Indonesia, political and religious leaders have decried the spread of hate speech and rumors over social media, which played a pivotal role in the Jakarta mayoral election. Mass media have been used to spread disinformation and hate speech in the past, and have played a key role in modern genocides. Social media is now playing a similar role in contemporary atrocities: in Burma, for instance, ultranationalist Buddhist monks have used social media to mobilize supporters and instigate violence against the Rohingya, a persecuted Muslim minority group. DISINFORMATION DURING ELECTIONS Often, disinformation aims to influence citizens’ decisions to vote (or to abstain from voting). The use of disinformation around elections is probably only slightly younger than representative democracy itself, but the reach, speed, and low cost of disseminating disinformation over social media has amplified this problem. The actors involved are often subnational political figures or organizations, although
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages7 Page
-
File Size-