Download the Report

Total Page:16

File Type:pdf, Size:1020Kb

Download the Report Ai in the Age of Cyber-Disorder Actors, Trends, and Prospects edited by Fabio Rugge © 2020 Ledizioni LediPublishing Via Antonio Boselli, 10 – 20136 Milan – Italy www.ledizioni.it [email protected] Ai In The Age Of Cyber-Disorder: Actors, Trends, And Prospects Edited by Fabio Rugge First edition: November 2020 Print ISBN 9788855263832 ePub ISBN 9788855263849 Pdf ISBN 9788855263856 DOI 10.14672/55263832 ISPI. Via Clerici, 5 20121, Milan www.ispionline.it Catalogue and reprints information: www.ledizioni.it This Report is realized with the support of the Policy Planning Unit of the Ministry of Foreign Affairs and International Cooperation pursuant to art. 23-bis of Presidential Decree 18/1967. The opinions contained in this Report are solely those of the authors and do not necessarily reflect the opinions of the Ministry of Foreign Affairs and International Cooperation, ISPI, and The Brookings Institution. The Report is produced within the framework of the activities of the Centre on Cybersecurity, jointly promoted by ISPI and Leonardo. The Brookings Institution is a nonprofit organization devoted to independent research and policy solutions. Its mission is to conduct high-quality, independent research and, based on that research, to provide innovative, practical recommendations for policymakers and the public. Table of Contents Introduction John R. Allen, Giampiero Massolo............................................ 8 1. AI in a Contested Cyberspace Fabio Rugge................................................................................... 12 Cybersecurity in AI National Strategies Thomas A. Campbell..................................................................56 2. Panopticon 2.0? AI Enabled Surveillance Practices in Authoritarian Regimes Samuele Dominioni.................................................................. 63 3. How AI Bots and Voice Assistants Reinforce Gender Bias Caitlin Chin, Mishaela Robison.......................................... 82 4. How To Deal with AI Enabled Disinformation? John Villasenor.......................................................................... 105 5. AI Revolution: Building Responsible Behavior Darrell M. West............................................................................ 118 6. Pacem in Cyberspace, Auspicia Algoretichs Paolo Benanti............................................................................. 134 About the Authors......................................................................... 154 Introduction In the summer of 1956, mathematician John McCarthy and another 19 scholars met at Dartmouth College on the outskirts of Hannover in New Hampshire for a summer research project on machine intelligence. This was the moment when what would later become known as Artificial Intelligence (AI) first emerged as a research discipline. Meanwhile, two other scholars, Harold J. Leavitt and Thomas L. Whisler, were writing an article speculating on new technologies and dubbing them information technology (IT). With this definition, they referred to technology of three types: for processing large amounts of information rapidly; for applying statistical and mathematical methods to decision-making problems; and for simulating higher-order thinking through computer programs. Artificial Intelligence and cyberspace were therefore born in the same historical context, and have been intertwined since their inception. Artificial Intelligence is a general-purpose technology that is considered one of the most promising of our times. To function, AI needs to be trained, and it finds its training material in data. These are collected, processed, exchanged and stored in the digital domain. According to a recent estimate, the global size of the data sphere reached 59 zettabytes in 2020. This is an unexpectedly large figure mainly due to the outbreak of the Covid-19 pandemic, which has pushed many around the world to consume and produce more digital data than last year. Indeed, a multitude of activities we used to do in-person have Introduction 9 turned into 2.0 happenings, including academic conferences, professional meetings, yoga classes, and even birthday parties. Recently, these have occurred in the digital domain (commonly called cyberspace) and generated an unprecedented amount of data. There is, nowadays, a sound literature investigating how AI could exploit data analysis to generate effects in our contemporary society, from medical applications to surveillance capitalism. In light of such understanding, cyberspace is one of the domains in which AI may play a role in achieving overarching objectives. For example, AI-enabled cyber-attacks may drive the next major upgrade in cyber-weapons, which could be used both by state and non-state actors. The relationship between AI and cyberspace has so far been investigated mostly in terms of the effects that AI could have on the digital domain and thus on our societies. What has been explored less is the opposite relationship, namely, how cyberspace geopolitics can affect AI. Therefore, in this second ISPI-Brookings Report, we decided to turn the picture upside-down and to focus our analysis on how ongoing dynamics in cyberspace as a whole may affect how AI technologies are developed and implemented. Indeed, cyberspace is not a neutral and aseptic environment. It is not even a utopian paradise where people can unite and meet across frontiers in a stateless digital reign. As Fabio Rugge argues in the opening chapter, cyberspace is now becoming the domain of choice for serious but ambiguous competition among nation states, and for destabilisation by non-state actors such as mercenaries, cybercriminals and hacktivists. This is the environment in which cyber-dependent technologies will be developed and fielded. As such, AI may accelerate the ongoing confrontation at the international level, adding yet another layer of unpredictability to an already troublesome and ambiguous digital domain. The chapters that follow will develop in two related directions. On the one hand, they will lead us to discover how an inadequately governed domain can have undesirable consequences for how AI technologies are 10 AI in the Age of Cyber-Disorder already implemented. On the other hand, they will also propose policies, and practical and ethical guidelines, to govern the use of AI more effectively and thus reduce the negative influence of geopolitical competition in cyberspace on AI applications. Indeed, the reckless use of AI can pose many risks and be detrimental to the founding principles of our international system, which is bound by the Universal Declaration of Human Rights. For example, the protection of human dignity at the individual level is currently endangered by the pervasiveness of surveillance programs, and particularly by those implemented in nondemocratic regimes. As Samuele Dominioni explains, authoritarian countries are exploiting the normative void around surveillance practices to strengthen their grip on power by building a kind of “digital panopticon”. The lack of proper regulations to establish pre-conditions for AI-enabled technologies is also evident in other seemingly domestic settings. As Caitlin Chin and Mishaela Robison demonstrate, AI bots and voice assistants reinforce gender bias due to societal tendencies to anthropomorphise objects by assigning gender. From offensive applications, pervasive surveillance, and algorithm biases, it seems that AI applications have suffered from a growing disorder and lack of enforceable normative solutions in the cyber domain. Yet, AI could be beneficial in tackling some of the threats currently coming from such a contested digital domain. As John Villasenor argues, AI could play a determining role in addressing the rapid and pervasive disinformation attacks that are increasingly affecting our societies. However, purely technological solutions are not enough. To tackle these challenges, actions must be multi-layered. On the one hand, as Darrell West argues, this era requires a “whole-of-society” and “whole-of-government” approach that considers how to maximise possible benefits and minimise costs. This includes creating uniform standards in terms of data access, data sharing, and data protection. On the other hand, we should never lose sight of the human dimension of technological developments. Along these lines, Paolo Benanti claims that it is necessary to Introduction 11 establish an international ethical governance system to identify and monitor developments in AI, weighing their potentially disruptive effects against their capabilities. As long as the cyber domain remains in turmoil, with no effective governance mechanism in place, it is likely that it will be challenging to find human-centric solutions that might allow humanity as a whole to benefit from the tremendous potential that AI has to offer. This report by ISPI and the Brookings Institution is an effort to shed light on this less studied, but extremely relevant, relationship. John R. Allen Giampiero Massolo President Brookings Institution President ISPI 1. AI in the Contested Cyberspace Fabio Rugge The Internet, born as an anarchic network to connect people across global frontiers, has become one of the most destabilizing areas of competition between states. The Internet is being militarized1, as cyberspace is now a domain where states’ power is projected, strategic goals are achieved with campaigns that fall under the definition of theuse of force, weapons systems are operated, wars will be fought. Cyberspace is also
Recommended publications
  • Disinformation in the Cyber Domain: Detection, Impact, and Counter-Strategies1
    24th International Command and Control Research and Technology Symposium ‘Managing Cyber Risk to Mission’ Disinformation in the Cyber Domain: Detection, Impact, and Counter-Strategies1 Ritu Gill, PhD Judith van de Kuijt, MA Defence R&D Canada Netherlands Organization-Applied Scientific Research [email protected] [email protected] Magnus Rosell, PhD Ronnie Johannson, PhD Swedish Defence Research Agency Swedish Defence Research Agency [email protected] [email protected] Abstract The authors examined disinformation via social media and its impact on target audiences by conducting interviews with Canadian Armed Forces and Royal Netherlands Army subject matter experts. Given the pervasiveness and effectiveness of disinformation employed by adversaries, particularly during major national events such as elections, the EU-Ukraine Association Agreement, and the Malaysian Airlines Flight 17, this study assessed several aspects of disinformation including i) how target audiences are vulnerable to disinformation, ii) which activities are affected by disinformation, iii) what are the indicators of disinformation, and iv) how to foster resilience to disinformation in the military and society. Qualitative analyses of results indicated that in order to effectively counter disinformation the focus needs to be on identifying the military’s core strategic narrative and reinforcing the larger narrative in all communications rather than continuously allocating valuable resources to actively refute all disinformation. Tactical messages that are disseminated should be focused on supporting the larger strategic narrative. In order to foster resilience to disinformation for target audiences, inoculation is key; inoculation can be attained through education as part of pre-deployment training for military, as well as public service announcements via traditional formats and through social media for the public, particularly during critical events such as national elections.
    [Show full text]
  • 2019 Brand Disinformation Impact Study January 2019
    2019 Brand Disinformation Impact Study January 2019 New Knowledge’s annual consumer study draws the connection between disinformation and the impact it has on brand reputation through the eyes of the everyday consumer. 2019 Brand Disinformation Impact Study What is the cost of trust in the digital age? This is the question that many brand executives ask themselves today as the way we consume and retain information evolves to keep up with technology. New Knowledge 2 2019 Brand Disinformation Impact Study But as more ways to access information arise, so too do the external forces that work to manipulate our information ecosystem, and make us question what is real and what is fake. New Knowledge 3 Introduction The threat of disinformation has New Knowledge recently partnered come front and center, as politics, with Survata to conduct an online institutions and brands face the consumer survey. The goal was to consequences of having few processes obtain a better understanding of what in place to defend information integrity consumers think about disinformation against false narratives and and its potential impact on highly- coordinated amplification. And the visible brand names. It was conducted impact is critical for brands, including in early January, 2019 with 1,500 brand reputation damage, and lost American consumers between the customers. Beyond the risk of ages of 18 and 65. The following is an in damaged brand reputation, is the depth look at the results. larger risk: loss of public trust. Trust cannot be bought or won. Brands know it takes transparency, and years of known, reputable work in the market to generate brand affinity.
    [Show full text]
  • Public Communication CHALLENGES
    30YEARS OF PUBLIC COMMUNICATION CHALLENGES Club of Venice 1986-2016 30YEARS OF PUBLIC COMMUNICATION CHALLENGES Club of Venice 1986-2016 TABLE OF CONTENT YEARS OF 30PUBLIC COMMUNICATION CHALLENGES This book was published in November 2016 on the occasion of the 30th anniversary of the Club of Venice. The contributions in this book reflect only the views of their respective authors. This publication has no commercial intent and cannot be used for any commercial purpose without prior agreement from its publishers. Published with the financial support of the Belgian State (Chancellery of the Prime Minister), the Government Offices of Sweden (Communications Division), the Government of the Italian Republic (Presidency of the Council of Ministers – Department for European Policies), the Ministry of General Affairs of the Netherlands (Public Information and Communication Office) and the European Parliament. Coordination: Philippe Caroyez, Hans Brunmayr, Vincenzo Le Voci Responsible editor: Belgian State – Chancellery of the Prime Minister Legal deposit: D/2016/9737/10 November 2016 Graphic design: Fabienne Bonnoron TABLE OF CONTENT FOREWORD Sandro Gozi 7 INTrodUCTION 8 Thirty years dedicated to a certain idea of Europe Stefano Rolando 8 The cardinal virtues of the Club of Venice... Philippe Caroyez 11 Fighting against fear and winning back confidence in Europe Aurelio Sahagún Pool 18 Sharing best practices in communicating on Europe for the sake of all its citizens Zvonimir Frka-Petešić 20 “Opening the space” and developing together Heike Thiele
    [Show full text]
  • Eeas Special Report Update: Short Assessment of Narratives and Disinformation Around the Covid-19 Pandemic (Update 23 April – 18 May)
    EEAS SPECIAL REPORT UPDATE: SHORT ASSESSMENT OF NARRATIVES AND DISINFORMATION AROUND THE COVID-19 PANDEMIC (UPDATE 23 APRIL – 18 MAY) Note: The objective of this report is to provide a snapshot overview of the current trends and insights into disinformation activities related to COVID-19. It does not provide a comprehensive or complete overview and focusses primarily on the external dimension, in line with the European External Action Service (EEAS) mandate. The report was authored by the EEAS Strategic Communications and Information Analysis Division, which contributes to the EU’s efforts to counter disinformation, including through detection, analysis and exposure of disinformation campaigns. In addressing disinformation and identifying and analyzing disinformation surrounding the COVID -19 outbreak, the EEAS is working closely with the European Commission, European Parliament services and EU Member States. The EEAS also cooperates on this issue with international partners (G7, NATO and non-state actors). The results are regularly published on the website of EEAS, EUvsDisinfo.eu and social media channels. Insights are shared with EU institutions and EU Member States in real time, including through the Rapid Alert System on Disinformation. SUMMARY In line with the previous EEAS Special Reports and based on publicly available information, this edition gives an overview of recent developments around COVID-19 related disinformation. The report focusses on foreign state and non-state actors trying to exploit the uncertainty in societies as well as people’s fears and insecurities to advance their own agenda. While misinformation and disinformation1 relating to issues surrounding the COVID-19 pandemic have continued to spread across the globe, the volume has – according to our findings – shown at least a temporary decrease during the period observed, alongside a general downward trend on COVID-19 related web searches2.
    [Show full text]
  • How Weaponizing Disinformation Can Bring Down a City's Power Grid
    How weaponizing disinformation can bring down a city's power grid Gururaghav Ramana, Bedoor AlSheblib,y, Marcin Waniekb,y, Talal Rahwanb,*, and Jimmy Chih-Hsien Penga,* aDepartment of Electrical and Computer Engineering, National University of Singapore, Singapore bComputer Science, New York University, Abu Dhabi, United Arab Emirates * Joint corresponding authors. E-mail: [email protected]; [email protected] yEqually contributing authors Abstract Social technologies have made it possible to propagate disinformation and manipulate the masses at an unprecedented scale. This is particularly alarming from a security perspective, as humans have proven to be the weakest link when protecting critical infrastructure in general, and the power grid in particular. Here, we consider an attack in which an adver- sary attempts to manipulate the behavior of energy consumers by sending fake discount notifications encouraging them to shift their consumption into the peak-demand period. We conduct surveys to assess the propen- sity of people to follow-through on such notifications and forward them to their friends. This allows us to model how the disinformation propagates through social networks. Finally, using Greater London as a case study, we show that disinformation can indeed be used to orchestrate an attack wherein unwitting consumers synchronize their energy-usage patterns, re- sulting in blackouts on a city-scale. These findings demonstrate that in an era when disinformation can be weaponized, system vulnerabilities arise arXiv:1908.02589v1 [cs.SI] 1 Aug 2019 not only from the hardware and software of critical infrastructure, but also from the behavior of the consumers. Introduction Social technologies have dramatically altered the ways in which conflicts are fought.
    [Show full text]
  • Disinformation Attack: How to Respond
    DISINFORMATION ATTACK: HOW TO RESPOND ANNELI AHONEN JANIS RUNGULIS EEAS EAST STRATCOM TASK FORCE EAST STRATCOM TASK FORCE MANDATE Communicating Strengthening Forecasting & OBJECTIVES EU policies media responding to effectively environment disinformation Go to EUvsDisinfo UNDERSTAND Disrupt public debate Undermine the Western order Exploit Project vulnerabilities worldview The information confrontation is part of Russian military doctrine and accepted by the top hierarchy of the leading Russian state-owned media "Word, camera, photo, internet: another type of weapon, another type of armed forces" Sergey Shoygu, Russian Minister of Defence RESOURCES Russia spends 1.3 billion EUR annually on official outlets 325 million EUR annually are given to Rossiya segodnya and RT RT operates in 100 countries Sputnik is available in more than 30 languages NARRATIVES The Elites vs. the People Threatened values ("Elites manipulate elections") ("Gays and lesbians issue dictates") Lost sovereignty Imminent collapse ("EU is occupied by the US") ("EU MS on the verge of civil war") HAHAGANDA HAHAGANDA HAHAGANDA HAHAGANDA HAHAGANDA MESSAGES IDENTIFY Hints HINTS Emotions Grammar URL Other articles Authors Activity CHECK European Commission: Euromyths European Parliament Think Tank: Disinformation Bellingcat DFR Lab IFCN The Insider Stopfake Polygraph EUvsDisinfo RESPOND SCENARIO You work for a local office of an international organization in a country of "strategic interest" to Russia. Your organization deals with healthcare issues. Recently your organization has been outspoken about the importance of vaccinations despite the growing resistance among local people. Monday. 9:30 Your local colleagues inform you that your organization has been mentioned in an article on a local fringe website 10:00 You check the website.
    [Show full text]
  • Disinformation in the Cyber Domain: Detection, Impact, and Counter-Strategies
    EasyChair Preprint № 1809 Disinformation in the Cyber Domain: Detection, Impact, and Counter-Strategies Ritu Gill, Judith van de Kuijt, Magnus Rossell and Ronnie Johansson EasyChair preprints are intended for rapid dissemination of research results and are integrated with the rest of EasyChair. October 31, 2019 24th International Command and Control Research and Technology Symposium ‘Managing Cyber Risk to Mission’ Disinformation in the Cyber Domain: Detection, Impact, and Counter-Strategies1 Ritu Gill, PhD Judith van de Kuijt, MA Defence R&D Canada Netherlands Organization-Applied Scientific Research [email protected] [email protected] Magnus Rosell, PhD Ronnie Johannson, PhD Swedish Defence Research Agency Swedish Defence Research Agency [email protected] [email protected] Abstract The authors examined disinformation via social media and its impact on target audiences by conducting interviews with Canadian Armed Forces and Royal Netherlands Army subject matter experts. Given the pervasiveness and effectiveness of disinformation employed by adversaries, particularly during major national events such as elections, the EU-Ukraine Association Agreement, and the Malaysian Airlines Flight 17, this study assessed several aspects of disinformation including i) how target audiences are vulnerable to disinformation, ii) which activities are affected by disinformation, iii) what are the indicators of disinformation, and iv) how to foster resilience to disinformation in the military and society. Qualitative analyses of results indicated that in order to effectively counter disinformation the focus needs to be on identifying the military’s core strategic narrative and reinforcing the larger narrative in all communications rather than continuously allocating valuable resources to actively refute all disinformation.
    [Show full text]
  • Traffic Networks Are Vulnerable to Disinformation Attacks
    www.nature.com/scientificreports OPEN Trafc networks are vulnerable to disinformation attacks Marcin Waniek1, Gururaghav Raman2, Bedoor AlShebli1,3, Jimmy Chih‑Hsien Peng2* & Talal Rahwan1* Disinformation continues to raise concerns due to its increasing threat to society. Nevertheless, a threat of a disinformation‑based attack on critical infrastructure is often overlooked. Here, we consider urban trafc networks and focus on fake information that manipulates drivers’ decisions to create congestion at a city scale. Specifcally, we consider two complementary scenarios, one where drivers are persuaded to move towards a given location, and another where they are persuaded to move away from it. We study the optimization problem faced by the adversary when choosing which streets to target to maximize disruption. We prove that fnding an optimal solution is computationally intractable, implying that the adversary has no choice but to settle for suboptimal heuristics. We analyze one such heuristic, and compare the cases when targets are spread across the city of Chicago vs. concentrated in its business district. Surprisingly, the latter results in more far‑reaching disruption, with its impact felt as far as 2 km from the closest target. Our fndings demonstrate that vulnerabilities in critical infrastructure may arise not only from hardware and software, but also from behavioral manipulation. Te ubiquity of social media and the internet has made it easier than ever to spread disinformation 1–4. Exacer- bating this phenomenon are the recent advances in machine learning and the rise of social bots, allowing (dis) information to be delivered to a target audience at an unprecedented scale 1,5,6.
    [Show full text]
  • Social Media Weaponization the Biohazard of Russian Disinformation Campaigns
    Senior Airman Marcel Williams, 27th Special Operations Wing public affairs broadcaster, speaks at “Gathering for Unity” event at Cannon Air Force Base, New Mexico, June 5, 2020, and shares experiencing racism in his own community (U.S. Air Force/Lane T. Plummer) Social Media Weaponization The Biohazard of Russian Disinformation Campaigns By Sarah Jacobs Gamberini n a renewed era of Great Power com- competitors are regularly operating allowed for nearly ubiquitous access to petition, the United States is faced below the threshold that would warrant its targets and a prolific capability for I with adversaries engaging across mul- a military response, including on the controlling a narrative and manipu- tiple domains without the traditional information battlefield. The blurred red lating the hearts and minds of a pop- distinctions of war and peace. America’s lines that result from covert informa- ulation on a range of sensitive societal tion operations waged by foreign actors issues, including public health. on the Internet will force a change in Russia has a long history of seeking how the United States operates and to project power and influence while Sarah Jacobs Gamberini is a Policy Fellow in how its society consumes information. playing with a technological and geopo- the Center for the Study of Weapons of Mass Destruction, Institute for National Strategic Russia used tactics of influence and litical handicap. Given its history and a Studies, at the National Defense University. coercion long before social media geographic location with many bordering 4 Forum / Social Media Weaponization JFQ 99, 4th Quarter 2020 nations, it sees itself as constantly be- encroachment from the United States region.9 Since the end of the Cold War, sieged from all sides, but particularly by and the North Atlantic Treaty Orga- Russia has had to be calculating and the West.
    [Show full text]
  • Russian Disinformation Attacks on Elections: Lessons from Europe
    RUSSIAN DISINFORMATION ATTACKS ON ELECTIONS: LESSONS FROM EUROPE HEARING BEFORE THE SUBCOMMITTEE ON EUROPE, EURASIA, ENERGY, AND THE ENVIRONMENT OF THE COMMITTEE ON FOREIGN AFFAIRS HOUSE OF REPRESENTATIVES ONE HUNDRED SIXTEENTH CONGRESS FIRST SESSION July 16, 2019 Serial No. 116–55 Printed for the use of the Committee on Foreign Affairs ( Available: http://www.foreignaffairs.house.gov/, http://docs.house.gov, or http://www.govinfo.gov U.S. GOVERNMENT PUBLISHING OFFICE 37–051PDF WASHINGTON : 2019 COMMITTEE ON FOREIGN AFFAIRS ELIOT L. ENGEL, New York, Chairman BRAD SHERMAN, California MICHAEL T. MCCAUL, Texas, Ranking GREGORY W. MEEKS, New York Member ALBIO SIRES, New Jersey CHRISTOPHER H. SMITH, New Jersey GERALD E. CONNOLLY, Virginia STEVE CHABOT, Ohio THEODORE E. DEUTCH, Florida JOE WILSON, South Carolina KAREN BASS, California SCOTT PERRY, Pennsylvania WILLIAM KEATING, Massachusetts TED S. YOHO, Florida DAVID CICILLINE, Rhode Island ADAM KINZINGER, Illinois AMI BERA, California LEE ZELDIN, New York JOAQUIN CASTRO, Texas JIM SENSENBRENNER, Wisconsin DINA TITUS, Nevada ANN WAGNER, Missouri ADRIANO ESPAILLAT, New York BRIAN MAST, Florida TED LIEU, California FRANCIS ROONEY, Florida SUSAN WILD, Pennsylvania BRIAN FITZPATRICK, Pennsylvania DEAN PHILLIPS, Minnesota JOHN CURTIS, Utah ILHAN OMAR, Minnesota KEN BUCK, Colorado COLIN ALLRED, Texas RON WRIGHT, Texas ANDY LEVIN, Michigan GUY RESCHENTHALER, Pennsylvania ABIGAIL SPANBERGER, Virginia TIM BURCHETT, Tennessee CHRISSY HOULAHAN, Pennsylvania GREG PENCE, Indiana TOM MALINOWSKI, New Jersey
    [Show full text]
  • Democratic Defense Against Disinformation 2.0
    Democratic Defense Against Disinformation 2.0 Alina Polyakova and Daniel Fried ISBN-13: 978-1-61977-592-3 This report is written and published in accordance with the Atlantic Council Policy on Intellectual Independence. The author is solely responsible for its analysis and recommendations. The Atlantic Council and its donors do not determine, nor do they necessarily endorse or advocate for, any of this report’s conclusions. June 2019 TABLE OF CONTENTS INTRODUCTION 1 Framing Solutions (Against a Moving Target) 1 PROGRESS REPORT 4 Europe Seeks Solutions 4 The G7 in the Game 6 The United States Tries to Get a Grip 7 Social Media Companies on the Spot 12 Civil Society Stays the Course 15 NEXT STEPS 16 The US Executive Should Get Serious 16 The US Congress Should Fill the Gap 20 Europe Should Make it Real 22 Social Media Companies Should Step Up 23 ABOUT THE AUTHORS 24 Democratic Defense Against Disinformation 2.0 INTRODUCTION his second edition of the paper, Democratic media companies, and civil-society groups invest in Defense Against Disinformation, seeks to cap- long-term resilience against disinformation, including ture the rapid development of policy responses raising social awareness of disinformation and encour- to the challenge—especially by governments aging digital literacy education, including how to dis- Tand social media companies—since initial publication cern fabricated or deceptive content and sources. in February 2018. The authors stand by the fundamen- tals of the earlier analysis and recommendations: that Democratic Defense Against Disinformation 2.0 will re- democratic societies can combat and mitigate the view developments over the past year, assess their ef- challenge of foreign disinformation while working with- fectiveness thus far, and offer suggestions for next steps.
    [Show full text]
  • Enterprise Appropriate Responses to Disinformation Risks
    705 5th AVE S | SUITE 800 SEATTLE, WA | 98104 206-800-1700 WWW.FIVEBY.COM Enterprise Appropriate Responses to Disinformation Risks 2021 Report Executive Summary Disinformation is a growing threat to communities and organizations, as it often manifests as changes in public narratives, beliefs, emotions, and actions. These changes are triggered by actors who use false, misleading or misattributed information with the intent to do harm, benefit financially or geopolitically, or damage social infrastructures. Disinformation campaigns exploit enterprise platforms, data, and communities almost daily; these, and organizations that confer credibility on domains, are part of both the campaigns and the potential defenses against them. Many of these organizations can monitor for and respond to relevant disinformation either themselves or through trusted third parties and groups. This report describes disinformation from a threat response perspective. We begin by describing the business aspects of disinformation, then address disinformation as an information security issue, providing solutions for adapting existing cybersecurity processes, tools, and teams to include “cognitive security” in their security enablement, operations, and test and validation work. We address specific areas, such as the data engineering needed to handle disinformation’s three Vs: Volume, Velocity, and Variety. We also tackle the adaptation of InfoSec standards, to allow rapid sharing of disinformation incident information to responders and the data governance issues involved in monitoring and patching systems whose boundaries are outside an information security team’s control. We cover cognitive security framed as a risk management activity and ways to predict, assess, reduce, and disrupt those risks, and we outline how disinformation response can be added either to an existing Security Operations Center (SOC) or organized as a standalone SOC.
    [Show full text]