India's Data Protection Framework Will Need to Treat Privacy As a Social and Not Just an Individual Good

Total Page:16

File Type:pdf, Size:1020Kb

India's Data Protection Framework Will Need to Treat Privacy As a Social and Not Just an Individual Good ISSN (Online) - 2349-8846 India's Data Protection Framework Will Need to Treat Privacy as a Social and Not Just an Individual Good AMBER SINHA Amber Sinha ([email protected]) is at the Centre of Internet and Society, and manages projects on privacy, big data and artificial intelligence. Vol. 53, Issue No. 18, 05 May, 2018 The idea that technological innovations may compete with privacy of individuals assumes that there is social and/or economic good in allowing unrestricted access to data. However, it must be remembered that data is potentially a toxic asset, if it is not collected, processed, secured and shared in the appropriate way. In July 2017, the Ministry of Electronics and Information Technology (MeITy) in India set up a committee headed by a former judge, B N Srikrishna, to address the growing clamour for privacy protections at a time when both private collection of data and public projects like Aadhaar are reported to pose major privacy risks (Maheshwari 2017). The Srikrishna Committee is in the process of providing its input, which will go on to inform India’s data- protection law. While the committee released a white paper with provisional views, seeking feedback a few months ago, it may be discussing a data protection framework without due consideration to how data practices have evolved. In early 2018, a series of stories based on investigative journalism by Guardian and Observer revealed that the data of 87 million Facebook users was used for the Trump campaign by a political consulting firm, Cambridge Analytica, without their ISSN (Online) - 2349-8846 permissions. Aleksandr Kogan, a psychology researcher at the University of Cambridge, created an application called “thisisyourdigitallife” and collected data from 270,000 participants through a personality test using Facebook’s application programming interface (API), which allows developers to integrate with various parts of the Facebook platform (Fruchter et al 2018). This data was collected purportedly for academic research purposes only. Kogan’s application also collected profile data from each of the participants’ friends, roughly 87 million people. The kinds of practices concerning the sharing and processing of data exhibited in this case are not unique. These are, in fact, common to the data economy in India as well. It can be argued that the Facebook–Cambridge Analytica incident is representative of data practices in the data-driven digital economy. These new practices pose important questions for data protection laws globally, and how these may need to evolve to address data protection, particularly for India, which is in the process of drafting its own data protection law. Privacy as Control Most modern data protection laws focus on individual control. In this context, the definition by the late Alan Westin (2015) characterises privacy as: The claim of individuals, groups, or institutions to determine for themselves when, how, and to what extent information about them is communicated to other. The idea of “privacy as control” is what finds articulation in data protection policies across jurisdictions, beginning with the Fair Information Practice Principles (FIPP) from the United States (US) (Dixon 2006). These FIPPs are the building blocks of modern information privacy law (Schwartz 1999) and not only play a significant role in the development of privacy laws in the US, but also inform data protection laws in most privacy regimes internationally (Rotenberg 2001), including the nine “National Privacy Principles” articulated by the Justice A P Shah Committee in India. Much of this approach is also reflected in the white paper released by the committee, led by Justice Srikrishna, towards the creation of data protection laws in India (Srikrishna 2017) This approach essentially involves the following steps (Cate 2006): (i) Data controllers are required to tell individuals what data they wish to collect and use and give them a choice to share the data. (ii) Upon sharing, the individuals have rights such as being granted access, and data controllers have obligations such as securing the data with appropriate technologies and procedures, and only using it for the purposes identified. The objective in this approach is to make the individual empowered and allow them to weigh ISSN (Online) - 2349-8846 their own interests in exercising their consent. The allure of this paradigm is that, in one elegant stroke, it seeks to “ensure that consent is informed and free and thereby also (seeks) to implement an acceptable tradeoff between privacy and competing concerns.” (Sloan and Warner 2014). This approach is also easy to enforce for both regulators and businesses. Data collectors and processors only need to ensure that they comply with their privacy policies, and can thus reduce their liability while, theoretically, consumers have the information required to exercise choice. In recent years, however, the emergence of big data, the “Internet of Things,” and algorithmic decision-making has significantly compromised the notice and consent model (Solove 2013). Limitations of Consent Some cognitive problems, such as long and difficult to understand privacy notices, have always existed with regard to the issue of informed consent, but lately these problems have become aggravated. Privacy notices often come in the form of long legal documents, much to the detriment of the readers’ ability to understand them. These policies are “long, complicated, full of jargon and change frequently” (Cranor 2012). Kent Walker (2001) lists five problems that privacy notices typically suffer from: (i) Overkill: Long and repetitive text in small print. (ii) Irrelevance: Describing situations of little concern to most consumers. (iii) Opacity: Broad terms that reflect limited truth, and are unhelpful to track and control the information collected and stored. (iv) Non-comparability: Simplification required to achieve comparability will lead to compromising of accuracy. (v) Inflexibility: Failure to keep pace with new business models. Today, data is collected continuously with every use of online services, making it humanly impossible to exercise meaningful consent. The quantity of data being generated is expanding at an exponential rate. With connected devices, smartphones, appliances transmitting data about our usage, and even the smart cities themselves, data now streams constantly from almost every sector and function of daily life, “creating countless new digital puddles, lakes, tributaries and oceans of information” (Bollier 2010). The infinitely complex nature of the data ecosystem renders consent of little value in cases where individuals may be able to read and comprehend privacy notices. As the uses of data are so diverse, and often not limited by a purpose identified at the beginning, individuals cannot conceptualise how their data will be aggregated and possibly used or reused. Seemingly innocuous bits of data revealed at different stages could be combined to reveal sensitive information about the individual. While the regulatory framework is designed such that individuals are expected to engage in cost–benefit analysis of trading their data to avail ISSN (Online) - 2349-8846 services, this ecosystem makes such individual analysis impossible. Conflicts Between Big Data and Individual Control The thrust of big data technologies is that the value of data resides not in its primary purposes, but in its numerous secondary purposes, where data is reused many times over (Schoenberger and Cukier 2013). On the other hand, the idea of privacy as control draws from the “data minimisation” principle, which requires organisations to limit the collection of personal data to the minimum extent necessary to obtain their legitimate purpose and to delete data no longer required. Control is excercised and privacy is enhanced by ensuring data minimisation. These two concepts are in direct conflict. Modern data-driven businesses want to retain as much data as possible for secondary uses. Since these secondary uses are, by their nature, unanticipated, their practices run counter to the very principle of purpose limitation (Tene and Polonetsky 2012). It is evident from such data-sharing practices, as demonstrated by the Cambridge Analytica–Facebook story, that platform architectures are designed with a clear view to collect as much data as possible. This is amply demonstrated by the provision of a “friends permission” feature by Facebook on its platform to allow individuals to share information not just about themselves, but also about their friends. For the principle of informed consent to be meaningfully implemented, it is necessary for users to have access to information about intended data practices, purposes and usage, so they consciously share data about themselves. In reality, however, privacy policies are more likely to serve as liability disclaimers for companies than any kind of guarantee of privacy for consumers. A case in point is Mark Zuckerberg’s facile claim that there was no “data-breach" in the Cambridge Analytica–Facebook incident. Instead of asking each of the 87 million users whether they wanted their data to be collected and shared further, Facebook designed a platform that required consent in any form only from 270,000 users. Not only were users denied the opportunity to give consent, their consent was assumed through a feature which was on by default. This is representative of how privacy trade-offs are conceived by current data- driven business models. Participation in a digital
Recommended publications
  • Information Warfare, International Law, and the Changing Battlefield
    ARTICLE INFORMATION WARFARE, INTERNATIONAL LAW, AND THE CHANGING BATTLEFIELD Dr. Waseem Ahmad Qureshi* ABSTRACT The advancement of technology in the contemporary era has facilitated the emergence of information warfare, which includes the deployment of information as a weapon against an adversary. This is done using a numBer of tactics such as the use of media and social media to spread propaganda and disinformation against an adversary as well as the adoption of software hacking techniques to spread viruses and malware into the strategically important computer systems of an adversary either to steal confidential data or to damage the adversary’s security system. Due to the intangible nature of the damage caused By the information warfare operations, it Becomes challenging for international law to regulate the information warfare operations. The unregulated nature of information operations allows information warfare to Be used effectively By states and nonstate actors to gain advantage over their adversaries. Information warfare also enhances the lethality of hyBrid warfare. Therefore, it is the need of the hour to arrange a new convention or devise a new set of rules to regulate the sphere of information warfare to avert the potential damage that it can cause to international peace and security. ABSTRACT ................................................................................................. 901 I. INTRODUCTION ......................................................................... 903 II. WHAT IS INFORMATION WARFARE? .............................
    [Show full text]
  • The Great Hack - Netflix Documentary Review
    The Great Hack - Netflix documentary review The Great Hack is a documentary discussing the ideas of how data and user actions on devices can create individual profiles of you and how companies can use these profiles and personality traits based on a user actions to target you on advertisements that are specifically for that type of character and drive persuasion. This is much more effective than randomly sending adverts to random users. The documentary focuses on a company that does this called Cambridge Analytica which is the worlds leading data driven communications company and David Carrol an associate professor who wokred on exposing this companies unethical and illegal ways of gaining access to users data and using it to drive votes for specific campaigns. Cambridge Analytica played a big part in the 2016 Presidential campaign. They spent 6 months sending surveys to users which were designed to create and find personailty profiles which were then sent back and would be targeted videos and other propaganda through media. This could be a video that would pop up on a persons recomended page on youtube which could be spreading shame to Hilary Clinton in this case and therefore drive the user to vote for Trump. Further on during the documentary, you are introduced to Carole Cadwalladr who was the investigative jounarlist for the Guardian. She investigated Cambridge Analytica and how it tied to the Brexit campaign. Other students should watch this as it gives you an incite on how big and important data can be and be used and manipulated and it one of the big reasons why actions like Brexit and Trump becoming president are in place today.
    [Show full text]
  • The Cambridge Analytica Files
    For more than a year we’ve been investigating Cambridge Analytica and its links to the Brexit Leave campaign in the UK and Team Trump in the US presidential election. Now, 28-year-old Christopher Wylie goes on the record to discuss his role in hijacking the profiles of millions of Facebook users in order to target the US electorate by Carole Cadwalladr Sun 18 Mar 2018 06:44 EDT The first time I met Christopher Wylie, he didn’t yet have pink hair. That comes later. As does his mission to rewind time. To put the genie back in the bottle. By the time I met him in person [www.theguardian.com/uk- news/video/2018/mar/17/cambridge-analytica- whistleblower-we-spent-1m-harvesting-millions-of- facebook-profiles-video], I’d already been talking to him on a daily basis for hours at a time. On the phone, he was clever, funny, bitchy, profound, intellectually ravenous, compelling. A master storyteller. A politicker. A data science nerd. Two months later, when he arrived in London from Canada, he was all those things in the flesh. And yet the flesh was impossibly young. He was 27 then (he’s 28 now), a fact that has always seemed glaringly at odds with what he has done. He may have played a pivotal role in the momentous political upheavals of 2016. At the very least, he played a From www.theguardian.com/news/2018/mar/17/data-war-whistleblower-christopher-wylie-faceook-nix-bannon- trump 1 20 March 2018 consequential role. At 24, he came up with an idea that led to the foundation of a company called Cambridge Analytica, a data analytics firm that went on to claim a major role in the Leave campaign for Britain’s EU membership referendum, and later became a key figure in digital operations during Donald Trump’s election [www.theguardian.com/us-news/2016/nov/09/how- did-donald-trump-win-analysis] campaign.
    [Show full text]
  • Ilnitnt ~Tares ~Cnetc JEFF FLAKE
    CHARLES E. GRASSLEY, IOWA, CHAIRMAN ORRIN G. HATCH, UTAH DIANNE FEINSTEIN. CALIFORNIA LINDSEY O GRAHAM, SOUTH CAROLINA PATRICK J LEAHY. VERMONT JOHN CORNYN, TEXAS RICHARD J DURBIN, ILLINOIS MICHAELS. LEE, UTAH SHELDON WHITEHOUSE, RHODE ISLAND TED CRUZ, TEXAS AMY KLOBUCHAR, MINNESOTA BEN SASSE. NEBRASKA CHRISTOPHER A . COONS, DELAWARE ilnitnt ~tares ~cnetc JEFF FLAKE. ARIZONA RICHARD BLUMENTHAL, CONNECTICUT MIKE CRAPO, IDAHO MAZIE K. HIRONO. HAWAII COMMITIEE ON THE JUDICIARY THOM TILLIS, NORTH CAROLINA CORY A. BOOKER, NEW JERSEY JOHN KENNEDY. LOUISIANA KAMALA D. HARRIS, CALIFORNIA WASHINGTON, DC 20510-6275 KoLAN L DAVIS, Ch,ef Counsel and Staff O,rector JENNIFER DUCK, Oemocrat,c Chief Counsel and Staff Director January 25, 2018 Mr. Stephen Bannon c/o William Burck Quinn Emanuel Urquhart & Sullivan, LLP 777 6th Street NW, I Ith floor Washington, D.C. 20001 Dear Mr. Bannon: As chief executive of the Trump campaign, a senior aide to the transition team, and former White House chief strategist, we believe that you have information that would assist the Committee in its investigation related to Russian interference in the 2016 presidential election and the dismissal of Director Corney. Accordingly, we are writing to request documents and schedule an interview with you in February 2018. Your positions placed you at key events of interest to the Committee in its investigation of Russian interference in the 2016 election and possible obstruction ofjustice . For example, it has been reported that you were involved in the Trump campaign's decision to hire Cambridge Analytica, 1 and were aware of WikiLeaks communications during the campaign with Donald Trump, Jr.
    [Show full text]
  • Cyberspace During a Crisis in Trust
    Governing Cyberspace during a Crisis in Trust An essay series on the economic potential — and vulnerability — of transformative technologies and cyber security Contents Governing Cyberspace during a Crisis in Trust . 4 Aaron Shull Tackling Cyber-enabled Crime Will Require Public-Private Leadership . 9 Neil Desai Election Cyber Security Challenges for Canada . 16 Elizabeth F. Judge and Michael Pal State and Surveillance . 21 David Lyon Trust and Data . 26 Paul Vallée Beware Fake News . 31 Eric Jardine The Need for a National Digital Identity Infrastructure . 36 Andre Boysen The Emerging Internet of Things . 41 Christopher S. Yoo The Cyber Security Battlefield . 45 Robert Fay and Wallace Trenholm TLD Operator Perspective on the Changing Cyber Security Landscape . 49 Byron Holland Strategic Stability, Cyber Operations and International Security . 55 David Mussington The Quantum Threat to Cyber Security . 60 Michele Mosca and Bill Munson Mitigating Cyber Risk across the Financial Sector . 64 Christian Leuprecht The Danger of Critical Infrastructure Interdependency . 69 Tyson Macaulay Programmable Trust . 74 Michael Mason and Matthew Spoke Patching Our Digital Future Is Unsustainable and Dangerous . 80 Melissa Hathaway Canada and Cyber Governance . 91 Stephanie Carvin Watch videos with series authors at cigionline.org/cyberspace Credits Executive Digital Media President Web Developer Rohinton P. Medhora Rebecca Anderson Interim Manager, Strategic Graphic Designer Initiatives and Special Projects Sami Chouhdary Liliana Araujo Multimedia Producer
    [Show full text]
  • Hacking the Electorate
    Hacking the Electorate On the Use of Personal Data in Political Campaigning www.kas.de Legal notice Publisher: Konrad-Adenauer-Stiftung e. V. 2020, Berlin Cover photo: © iStock/Orbon Alija Chapter marker: p. 10 © Adobe Stock/Gorodenkoff; p. 28 © Adobe Stock/Alexander; p. 38 © Shutterstock/mrmohock Design and typesetting: yellow too, Pasiek Horntrich GbR The print edition of this publication was printed by copy print Kopie & Druck GmbH, Berlin. Printed in Germany. Produced with financial support from the Federal Republic of Germany. The text of this publication is licensed under the terms of “Creative Commons Attribution-ShareAlike 4.0 International”, CC BY-SA 4.0 (available at: https://creativecommons.org/licenses/by-sa/4.0/legalcode.de). ISBN 978-3-95721-772-1 Hacking the Electorate On the Use of Personal Data in Political Campaigning At a Glance › Although data-driven political campaigning is not a new phenomenon, the tools used, the amount of data accessible, and the potential capacity to influence voters represent a new and challenging scenario for the rule of law. › With the arrival of participatory and social web, Internet users can now generate data in a complex network and without any obligation to the pursuit of objectivity or journalistic standards as pillars for content creation. › People in different countries are increasingly getting informed and learning about political candidates and other political issues through social networks. › In recent years political parties and campaigners around the world have invested heavily in online advertising, demonstrating all the potential to reach more people in an efficient, targeted, and accessible way.
    [Show full text]
  • Ad Hoc Session of the United Nations Commission on Science and Technology
    Ad Hoc Session of The United Nations Commission on Science and Technology Specialized Agency Queen’s National Model United Nations 2021 1 International Tech Summit 2018 From the Secretary-General Dear Delegates, It is my privilege to welcome you to Queen’s National Model United Nations 2021! My name is Samuel Tobert, and it is my honour to serve as your Secretary-General for QNMUN 2021. Originally from Vancouver, B.C. I study Biochemistry, Political Science and Business at Queen’s University. Model UN has been an integral part of my High School and University careers, and through it, I have met some of my closest friends, made some of my fondest memories, and learned valuable skills, becoming a more engaged global citizen. I am so glad that you have chosen to join us for this year’s iteration of QNMUN, held virtually for the first time in our history. COVID-19 has brought many challenges to our community and has presented new obstacles to Model UN; however, I can confidently say that our team has been working tirelessly to deliver a high quality and engaging conference for you. If there is anything we can do to make our conference more accessible in its online format, please do not hesitate to reach out. With this year marking my 8th on the Model UN circuit, I reflect upon the many experiences that I have had in MUN. I can confidently say that Model UN is one of the most fulfilling activities I have had the privilege to participate in. It has made me a better speaker, a better problem solver and taught me so much about myself and the world we live in.
    [Show full text]
  • Twitter Analyzer—How to Use Semantic Analysis to Retrieve an Atmospheric Image Around Political Topics in Twitter
    big data and cognitive computing Article Twitter Analyzer—How to Use Semantic Analysis to Retrieve an Atmospheric Image around Political Topics in Twitter Stefan Spettel 1 and Dimitrios Vagianos 2,* 1 Faculty of Informatics, Technical University of Vienna, Karlsplatz 13, 1040 Wien, Austria 2 Department of International and European Studies, University of Macedonia, Egnatia 156, 54636 Thessaloniki, Greece * Correspondence: [email protected]; Tel.: +30-2310-891-487; Fax: +30-2310-891-285 Received: 5 June 2019; Accepted: 2 July 2019; Published: 6 July 2019 Abstract: Social media are heavily used to shape political discussions. Thus, it is valuable for corporations and political parties to be able to analyze the content of those discussions. This is exemplified by the work of Cambridge Analytica, in support of the 2016 presidential campaign of Donald Trump. One of the most straightforward metrics is the sentiment of a message, whether it is considered as positive or negative. There are many commercial and/or closed-source tools available which make it possible to analyze social media data, including sentiment analysis (SA). However, to our knowledge, not many publicly available tools have been developed that allow for analyzing social media data and help researchers around the world to enter this quickly expanding field of study. In this paper, we provide a thorough description of implementing a tool that can be used for performing sentiment analysis on tweets. In an effort to underline the necessity for open tools and additional monitoring on the Twittersphere, we propose an implementation model based exclusively on publicly available open-source software. The resulting tool is capable of downloading Tweets in real-time based on hashtags or account names and stores the sentiment for replies to specific tweets.
    [Show full text]
  • Supplementary Evidence from Chris Wylie
    A RESPONSE TO MISSTATEMENTS IN RELATION TO CAMBRIDGE ANALYTICA INTRODUCTORY BACKGROUND TO THE COMPANIES Mr Wylie was the Director of Research for SCL and Cambridge Analytica from 2013 to the end of 2014. SCL Group is a UK-based military contractor that specialises in Information Operations (“IO”). SCL’s clients have included the UK Ministry of Defence, US Department of Defense and various NATO militaries. Information Operations is the area of military strategy that deploys, manipulates or weaponises information to support operational objectives. Within IO, there are related fields such as Psychological Operations and Cyber Operations. It is important to highlight that as IO is a military strategy, which is often deployed in combat situations where the Data Protection Act would not apply, many IO approaches are not generally congruent with the Data Protection Principles. This is because there are two key objectives of IO. The first is the notion of “informational dominance”, which focuses on capturing, interfering or manipulating as many channels of information surrounding the target as possible. This is typically done, by necessity, without the knowledge of the target. The second is using information collected about the target to identify and then exploit mental vulnerabilities to provoke certain behaviours in the target that would be conducive to operational objectives. Cambridge Analytica (“CA”) was created by SCL Group with funding from Robert Mercer, an American billionaire based in New York. Robert Mercer installed the alt-right political activist Stephen Bannon as CA’s Vice President with responsibilities to manage the company day-to-day. Mr Mercer wanted to use the IO tactics SCL had used on military projects for his political aims in the United States, and elsewhere, including the United Kingdom.
    [Show full text]
  • Disinformation and 'Fake News': Interim Report
    House of Commons Digital, Culture, Media and Sport Committee Disinformation and ‘fake news’: Interim Report Fifth Report of Session 2017–19 Report, together with formal minutes relating to the report Ordered by the House of Commons to be printed 24 July 2018 HC 363 Published on 29 July 2018 by authority of the House of Commons The Digital, Culture, Media and Sport Committee The Digital, Culture, Media and Sport Committee is appointed by the House of Commons to examine the expenditure, administration and policy of the Department for Digital, Culture, Media and Sport and its associated public bodies. Current membership Damian Collins MP (Conservative, Folkestone and Hythe) (Chair) Clive Efford MP (Labour, Eltham) Julie Elliott MP (Labour, Sunderland Central) Paul Farrelly MP (Labour, Newcastle-under-Lyme) Simon Hart MP (Conservative, Carmarthen West and South Pembrokeshire) Julian Knight MP (Conservative, Solihull) Ian C. Lucas MP (Labour, Wrexham) Brendan O’Hara MP (Scottish National Party, Argyll and Bute) Rebecca Pow MP (Conservative, Taunton Deane) Jo Stevens MP (Labour, Cardiff Central) Giles Watling MP (Conservative, Clacton) The following Members were also members of the Committee during the inquiry Christian Matheson MP (Labour, City of Chester) Powers The Committee is one of the departmental select committees, the powers of which are set out in House of Commons Standing Orders, principally in SO No 152. These are available on the internet via www.parliament.uk. Publication Committee reports are published on the Committee’s website at www.parliament.uk/dcmscom and in print by Order of the House. Evidence relating to this report is published on the inquiry publications page of the Committee’s website.
    [Show full text]
  • Using Care Ethics to Analyze the Facebook-Cambridge Analytica Scandal
    Using Care Ethics to Analyze the Facebook-Cambridge Analytica Scandal STS Research Paper Presented to the Faculty of the School of Engineering and Applied Science University of Virginia By Luke Deni March 1, 2020 On my honor as a University student, I have neither given nor received unauthorized aid on this assignment as defined by the Honor Guidelines for Thesis-Related Assignments. Signed: _______________________________________________ Approved: _______________________________________ Date ________________________ Benjamin J. Laugelli, Assistant Professor, Department of Engineering and Society Introduction In March of 2018, news broke that Cambridge Analytica was able to gather Facebook profile data of over 79 million users. They used this data to build voter personality profiles to target advertisements for the conservative presidential candidates in the 2016 election. Aleksandr Kogan, a social psychologist at the University of Cambridge, developed a third party application that 300,000 Facebook users downloaded, allowing him to collect information on their activity, their interests, and also the information of their friends, which he illegally transferred to Cambridge Analytica. Facebook denied that it was a data breach because users technically agreed on their privacy settings. This resulted in a public apology, claiming they will do better to protect user’s information (Kozlowska, 2018). The Facebook-Cambridge Analytica Scandal has become a large debate about privacy on the growing internet. There has been little discussion, however, of the issue of transparency between Facebook and its users. This is due to the large focus on the numbers itself: 79 million profiles. Despite popular belief, Facebook does offer privacy features through its user settings. The issue is they are optional features that default to loose restrictions which are also hard to find (Kozlowska, 2018).
    [Show full text]
  • The Great Hack Worksheet B1
    The Great Hack worksheet B1 The Great Hack documentary uncovers the different sides of the Cambridge Analytica/ Facebook data scandal. Award-winning filmmakers Karim Amer and Jehane Noujaim discuss their motivation to make a film about the origin of the information we consume daily and access to the personal data we share on the Internet and social media. Professor David Carroll teaches social-media design at Parsons School in New York. He decided to attack Cambridge Analytica in British courts, asking to see the data they held on him. Photos and interview: Netflix www.speakeasy-news.com - August 2019 B1 The Great Hack worksheet 1 Brittany Kaiser was the Cambridge Analytica’s business development director. She has taken on the role of whistleblower, giving detailed information to the courts, British Parliament and US Congress committees on Cambridge Analytica’s practices. Carole Cadwalladr is a British investigative journalist who broke the stories that Cambridge Analytica may have used data obtained from Facebook to try to manipulate voters in 2016 in the Presidential elections in the U.S.A. and Brexit referendum in the U.K. Photos and interview: Netflix www.speakeasy-news.com - August 2019 B1 The Great Hack worksheet 2 Q&A WITH KARIM AMER AND JEHANE NOUJAIM 1 What inspired you to make The Great Hack? For us a hack is about finding a vulnerability in the system and exploiting it. We realized that the true hack of today’s world was the hack of the mind. All the tiny manipulations that can happen through a screen and begin to warp where we go, what we do, and who we are.
    [Show full text]