Platform Liability in the EU: a Need for Reform? 2020

Total Page:16

File Type:pdf, Size:1020Kb

Platform Liability in the EU: a Need for Reform? 2020 Platform Liability in the EU: A Need for Reform? 2020 3(53) Platform Liability in the EU: A Need for Reform? Table of Contents Definitions ................................................................................................. 4 1 Executive summary ........................................................................... 5 2 Introduction ....................................................................................... 7 2.1 The purpose of the report ............................................................ 8 2.2 Method ........................................................................................ 8 2.3 Overview ..................................................................................... 8 2.4 The context of the report ............................................................. 9 2.4.1 Platforms and the internal market ....................................... 9 2.4.2 Platforms and sustainable development ............................ 10 3 The regulatory framework for platform content liability in the EU 11 3.1 The e-Commerce Directive ....................................................... 11 3.2 The Copyright Directive (2019) ............................................... 14 3.3 The Audiovisual Media Services Directive (2018) .................. 15 3.4 The relationship and consistency between the directives ......... 16 3.5 Self-regulation .......................................................................... 17 4 A further harmonised framework for platform liability in the EU .. 19 4.1 Evidence of fragmentation between the Member States .......... 19 4.2 A harmonised framework for platform liability: issues for further analysis .................................................................................... 24 5 A stricter level of platform liability in the EU? ............................... 25 5.1 Main arguments in favour of a stricter liability ........................ 26 5.2 Main arguments against a stricter liability ................................ 28 5.3 Recommendations ..................................................................... 31 Bibliography and other sources ............................................................... 33 Sammanfattning på svenska (Summary in Swedish) .............................. 39 Annex: Platform Liability and the e-Commerce Directive – answers from stakeholders ............................................................................................. 41 4(53) Definitions The following definitions apply for the purpose of this report. Platforms: A platform is defined as a “digital service that facilitates interactions between two or more distinct but interdependent sets of users (whether firms or individuals) who interact through the service via the internet”.1 The platform business model is characterised by the fact that platforms act as intermediaries between two or more types of users.2 Examples include e-commerce platforms that match retailers and customers, social media platforms that match users and advertisers and apartment sharing platforms that match apartment owners with customers seeking short-term rentals. Illegal content: Illegal content is defined as content that infringes European Union (EU) law or national laws in the Member States,3 including illegal offers for sale of goods and services. Falling outside the scope of the report are discussions regarding content that is not illegal but rather “harmful”, such as pornographic content. The distinction between illegal and harmful content is of course fluid and changes subject to legislative developments. To take one example, the concept of “fake news” today exists in the grey zone between legal and illegal content. Some countries are taking measures to classify “fake news” as illegal,4 a trend that has intensified during the COVID-19 pandemic.5 In this report, we will not discuss “fake news” as an issue of illegal content. This is in line with the approach taken by the Commission in the recent consultation on the Digital Services Act.6 Liability: When discussing the liability of platforms in this report, we are addressing under what circumstances platforms can or cannot be held liable for the content posted on the platform by its users. This is commonly referred to as the secondary liability of platforms. If the platform were to publish illegal content on its own, this would concern the primary liability of platforms. The primary liability of platforms falls beyond the scope of this report. Content filtering: When discussing content filtering in this report we are referring to the use of software and algorithms to prevent, screen and remove illegal content from platforms. A common example is the ContentID system used by YouTube, which is a digital fingerprinting system to detect copyright infringements. Content filtering in a wider sense may also refer to measures taken by other actors. Examples include an internet access provider that blocks access to certain websites or a parent who uses software to block certain internet content for children. The latter issues are however not the subject of this paper. 1 OECD (2019) p. 21. 2 See Stefan Larsson & Jonas Andersson Schwarz (2018), p. 20. 3 This is also the definition used in the recommendation by the Commission on measures to tackle illegal content online. See C(2018) 1177. 4 Michael-Ross Fiorentino (2018). 5 EDRi (2020). 6 Commission (2020a). 5(53) 1 Executive summary This report discusses the need to reform the European Union (EU) rules on platform liability for illegal content, such as terrorist propaganda, hate speech and illegal products. In this regard, the President of the European Commission Ursula von der Leyen has called for a Digital Services Act that will upgrade the liability and safety rules for platforms in the EU. The purpose of this report is to outline key issues and arguments that need to be considered in the upcoming policy process. The report forms part of the work of the National Board of Trade in offering analysis to the Swedish government ahead of upcoming negotiations regarding the internal market. The analysis presented in this report is supported by the results obtained through a questionnaire sent out by the National Board of Trade to stakeholders in December 2019 (the NBT questionnaire). The questionnaire was answered by 29 stakeholders and will be used to provide a sample of their insights. The questionnaire will not be used to draw statistical conclusions due to the relatively small sample size. The answers to the questionnaire are summarised in a document and can be found as a separate annex to this report. The analysis in this report is divided into two parts. The first part discusses fragmentation under the current regulatory framework and potential improvements to increase harmonisation. The second part discusses the arguments for and against a more large-scale reform of the level of liability for platforms. The discussion in this latter part focuses on the need for the introduction of mandatory filtering obligations. Analysis In terms of harmonisation under the current regulatory framework, we present evidence of fragmentation between the Member States. Indeed, the national regulations have to a certain extent diverged over the 20 years since the e-Commerce Directive was enacted. To address this issue, we highlight the following issues for further analysis: • The possibility of introducing a harmonised notice and takedown procedure across the EU, including issues such as the territorial scope of takedowns and access to justice. • The possibility of reforming the scope of the e-Commerce Directive to account for “new services” that were not explicitly addressed in the original directive. 6(53) • The possibility of clarifying the distinction between active and passive hosting services. In this regard, the legislator may consider a clarification that good faith efforts from platforms to prevent illegal content will not lead to the loss of the liability exception. • The possibility of clarifying the distinction between general and specific monitoring obligations, in line with recent case law. • The possibility of clarifying the relationship between the e- Commerce Directive and other pieces of EU platform legislation, such as the Copyright Directive and the AVMS Directive. With regard to the level of platform liability, there are forceful arguments for and against a reform. The situation is characterised by a complex balancing act between competing interests. This was also reflected in the answers to the NBT questionnaire, where stakeholders were divided as to the need for a large-scale reform of the rules on hosting liability. We present the following recommendations for the policy process moving forward. First of all we stress the importance of gathering solid evidence. It is crucial that the policy process is not rushed and that proposals are based on solid empirical evidence. Examples of tools that can be used to gather evidence include regulatory sandboxes and public consultations. Second, we note the fact that illegal content is a multi-faceted problem. Some areas of content liability may be more suitable for reform than others. For example, stricter liability might be investigated with regard to child sexual abuse material and illegal goods. Finally, it is important that any reform respects fundamental freedoms, such as the freedom of expression and the freedom to conduct a business. 7(53) 2 Introduction This year marks the 20th anniversary of the e-Commerce Directive. The Directive
Recommended publications
  • Reform of the EU Liability Regime for Online Intermediaries
    Reform of the EU liability regime for online intermediaries Background on the forthcoming digital services act IN-DEPTH ANALYSIS EPRS | European Parliamentary Research Service Author: Tambiama Madiega Members' Research Service PE 649.404 – May 2020 EN The E-commerce Directive, adopted in 2000, aims at harmonising minimum standards of liability for internet (online) intermediaries across the European Union (EU). Under the current EU legal framework, digital or online platforms are not legally responsible for hosting illegal content, but are required to remove such material once it is flagged. However, technologies and business models have evolved in the last 20 years and the EU liability framework struggles to capture liability issues raised by new actors, such as search engines, social networks and online marketplaces. Furthermore, societal challenges have changed the nature and scale of liability, with the development of a range of new harmful online practices, such as the dissemination of terrorist content online, the increasing use of platforms to distribute counterfeit products and the spreading of false or misleading news and online advertisements. The European Commission has pledged to present a new framework to increase and harmonise the responsibilities of online platforms. New EU legislation is expected in the forthcoming digital services act package, and this will set rules on how companies such as Google, Facebook and Twitter will have to police illegal and possibly harmful content online. Against this background, this paper aims to (1) describe the current EU liability regime for online intermediaries set out under the E-commerce Directive; (2) highlight the implementation gaps that have been identified; and (3) present the main proposals for reforming such an online liability regime and the main policy questions that have been discussed so far.
    [Show full text]
  • Regulation of the Digital Sector
    EU-US EXPLAINER July 2021 Regulation of the digital sector With online platforms and markets enmeshed in our societies and economies, the need to revisit and update existing digital regulations is becoming increasingly apparent. The debate around these reforms in the US, the EU and elsewhere touches on fundamental questions of privacy, transparency and free speech and the dynamic between private firms and governmental oversight is complex. While online platforms play a salient role in daily life, both the US and the EU continue to operate with regulations dating back over a generation. As significant challenges regarding illegal and harmful online content and moderation liability continue to have real world effects today, both the EU and the US are currently considering precedent-setting updates. Background: Existing EU and US digital regulations The close trading ties between the US and the EU in the area of digitally enabled services have been key to the growth of the digital economy. With 450 million users, the EU is one of the most profitable markets for US tech firms. This has enabled the US to export information and communications technology (ICT) services worth US$31 billion to the EU, with ICT-enabled services potentially adding another US$196 billion. Digital regulation has proved challenging, however, as the seminal laws governing conduct and liability of online platforms in both the EU and US are over 20 years old, with few updates over the years. In the US, a crucial piece of governing legislation is Section 230 of the Communications Decency Act, passed in 1996. Section 230 consists of two primary elements: the first removing liability for third-party content on platforms and the second encouraging platforms to self-moderate.
    [Show full text]
  • How Law Made Silicon Valley
    Emory Law Journal Volume 63 Issue 3 2014 How Law Made Silicon Valley Anupam Chander Follow this and additional works at: https://scholarlycommons.law.emory.edu/elj Recommended Citation Anupam Chander, How Law Made Silicon Valley, 63 Emory L. J. 639 (2014). Available at: https://scholarlycommons.law.emory.edu/elj/vol63/iss3/3 This Article is brought to you for free and open access by the Journals at Emory Law Scholarly Commons. It has been accepted for inclusion in Emory Law Journal by an authorized editor of Emory Law Scholarly Commons. For more information, please contact [email protected]. CHANDER GALLEYSPROOFS2 2/17/2014 9:02 AM HOW LAW MADE SILICON VALLEY Anupam Chander* ABSTRACT Explanations for the success of Silicon Valley focus on the confluence of capital and education. In this Article, I put forward a new explanation, one that better elucidates the rise of Silicon Valley as a global trader. Just as nineteenth-century American judges altered the common law in order to subsidize industrial development, American judges and legislators altered the law at the turn of the Millennium to promote the development of Internet enterprise. Europe and Asia, by contrast, imposed strict intermediary liability regimes, inflexible intellectual property rules, and strong privacy constraints, impeding local Internet entrepreneurs. This study challenges the conventional wisdom that holds that strong intellectual property rights undergird innovation. While American law favored both commerce and speech enabled by this new medium, European and Asian jurisdictions attended more to the risks to intellectual property rights holders and, to a lesser extent, ordinary individuals.
    [Show full text]
  • European Union Update: the Digital Services Act
    European Union Update: The Digital Services Act ICANN Government & Intergovernmental Organization (IGO) Engagement Elena Plexida GE-004 15 July 2020 ICANN | European Union Update: The Digital Services Act | July 2020 | 1 TABLE OF CONTENTS Foreword 3 What is the Digital Services Act? 4 Why a DSA? 4 What will the DSA likely contain? 5 How are ICANN and the ICANN community affected? 6 Next steps 6 ICANN | European Union Update: The Digital Services Act | July 2020 | 2 Foreword This paper provides an update and analysis of the Digital Services Act (DSA) initiative launched by the European Union (EU). The DSA initiative was identified in the first of the series of the ICANN Government & Intergovernmental Organization (IGO) Engagement papers capturing EU policy updates as an area of potential relevance to the ICANN community and to ICANN’s technical role in the Internet ecosystem.1 This paper broadly explains the initiative to date and its potential impact. It is worth noting that the European Commission launched a public consultation on the DSA initiative that will remain open until 8 September 2020.2 ICANN org will submit a contribution to the open consultation and encourages the community to take this opportunity to contribute as well. ICANN organization will continue to monitor the DSA initiative and will provide updates as relevant developments unfold. 1 https://www.icann.org/en/system/files/files/eu-policy-update-03apr20-en.pdf 2 https://ec.europa.eu/digital-single-market/en/news/consultation-digital-services-act- package ICANN | European
    [Show full text]
  • No Monitoring Obligations’ the Death of ‘No Monitoring Obligations’ a Story of Untameable Monsters by Giancarlo F
    The Death of ‘No Monitoring Obligations’ The Death of ‘No Monitoring Obligations’ A Story of Untameable Monsters by Giancarlo F. Frosio* Abstract: In imposing a strict liability regime pean Commission, would like to introduce filtering for alleged copyright infringement occurring on You- obligations for intermediaries in both copyright and Tube, Justice Salomão of the Brazilian Superior Tribu- AVMS legislations. Meanwhile, online platforms have nal de Justiça stated that “if Google created an ‘un- already set up miscellaneous filtering schemes on a tameable monster,’ it should be the only one charged voluntary basis. In this paper, I suggest that we are with any disastrous consequences generated by the witnessing the death of “no monitoring obligations,” lack of control of the users of its websites.” In order a well-marked trend in intermediary liability policy to tame the monster, the Brazilian Superior Court that can be contextualized within the emergence of had to impose monitoring obligations on Youtube; a broader move towards private enforcement online this was not an isolated case. Proactive monitoring and intermediaries’ self-intervention. In addition, fil- and filtering found their way into the legal system as tering and monitoring will be dealt almost exclusively a privileged enforcement strategy through legisla- through automatic infringement assessment sys- tion, judicial decisions, and private ordering. In multi- tems. Due process and fundamental guarantees get ple jurisdictions, recent case law has imposed proac- mauled by algorithmic enforcement, which might fi- tive monitoring obligations on intermediaries across nally slay “no monitoring obligations” and fundamen- the entire spectrum of intermediary liability subject tal rights online, together with the untameable mon- matters.
    [Show full text]
  • Response to the European Commission's
    Response to the European Commission’s “Inception Impact Assessment for an Ex-Ante Regulatory Instrument of Very Large Online Platforms Acting as Gatekeepers” INTRODUCTION On behalf of the Center for Data Innovation (datainnovation.org), we are pleased to submit feedback to the European Commission’s roadmap titled “Digital Services Act package: Ex-ante regulatory instrument of very large online platforms acting as gatekeepers.”1 In this submission, we discuss a number of assumptions and claims the Commission's proposal is based on, analyze the consequences of this proposal, and identify recommendations the Commission should consider. The Center for Data Innovation is the leading think tank studying the intersection of data, technology, and public policy. With staff in Washington, D.C. and Brussels, the Center formulates and promotes pragmatic public policies designed to maximize the benefits of data-driven innovation in the public and private sectors. It educates policymakers and the public about the opportunities and challenges associated with data, as well as technology trends such as open data, cloud computing, and the Internet of Things. The Center is a non-profit, non-partisan research institute affiliated with the Information Technology and Innovation Foundation (ITIF). BACKGROUND In recent years, a number of discursive strategies, laws, and rules have reflected the growing sense among EU policymakers that digital platforms have been under regulated. In the EU, there is an ongoing debate as to if and how much the high degree of market
    [Show full text]
  • Regulating Disinformation in Europe: Implications for Speech and Privacy
    UC Irvine Journal of International, Transnational, and Comparative Law Volume 6 Symposium: The Transnational Legal Ordering of Privacy and Speech Article 3 5-27-2021 Regulating Disinformation in Europe: Implications for Speech and Privacy Joris van Hoboken Vrije Universiteit Brussel Ronan Ó Fathaigh University of Amsterdam Follow this and additional works at: https://scholarship.law.uci.edu/ucijil Part of the Comparative and Foreign Law Commons, International Law Commons, Internet Law Commons, Privacy Law Commons, and the Transnational Law Commons Recommended Citation Joris v. Hoboken & Ronan Ó Fathaigh, Regulating Disinformation in Europe: Implications for Speech and Privacy, 6 UC Irvine Journal of International, Transnational, and Comparative Law 9 (2021). Available at: https://scholarship.law.uci.edu/ucijil/vol6/iss1/3 This Article is brought to you for free and open access by UCI Law Scholarly Commons. It has been accepted for inclusion in UC Irvine Journal of International, Transnational, and Comparative Law by an authorized editor of UCI Law Scholarly Commons. Regulating Disinformation in Europe: Implications for Speech and Privacy Joris van Hoboken* & Ronan Ó Fathaigh** This Article examines the ongoing dynamics in the regulation of disinformation in Europe, focusing on the intersection between the right to freedom of expression and the right to privacy. Importantly, there has been a recent wave of regulatory measures and other forms of pressure on online platforms to tackle disinformation in Europe. These measures play out in different ways at the intersection of the right to freedom of expression and the right to privacy. Crucially, as governments, journalists, and researchers seek greater transparency and access to information from online platforms to evaluate their impact on the health of their democracies, these measures raise acute issues related to user privacy.
    [Show full text]
  • Draft Progress Report
    Council of the European Union Brussels, 6 May 2021 (OR. en) 8415/21 Interinstitutional File: 2020/0361(COD) COMPET 318 MI 306 JAI 477 TELECOM 179 CT 57 PI 31 AUDIO 46 CONSOM 107 CODEC 637 IA 79 NOTE From: Presidency/General Secretariat of the Council To: Permanent Representatives Committee No. Cion doc.: 14124/20 + COR 1 + ADD 1 Subject: Proposal for a Regulation of the European Parliament and of the Council on a Single Market for Digital Services (Digital Services Act) and amending Directive 2000/31/EC - Draft progress report I. INTRODUCTION 1. On 15 December 2020, the Commission submitted the above-mentioned proposal for a Regulation of the European Parliament and the Council1. The proposal is based on Article 114 of the Treaty on the Functioning of the European Union (TFEU). 2. The proposal for a Regulation aims to contribute to the proper functioning of the internal market for intermediary services by setting out uniform rules for a safe, predictable and trusted online environment, where fundamental rights enshrined in the Charter are effectively protected. 1 Doc. 14124/20 + COR 1 + ADD 1. 8415/21 US/TM/sk 1 ECOMP 3.A EN 3. The European Economic and Social Committee provided its opinion on the proposal on 27 April 2021.2 4. In the European Parliament, the Committee on the Internal Market and Consumer Protection (IMCO) has not yet voted on its report. 5. In the statement of 25 March 2021, the members of the European Council invited the co- legislators to work swiftly on the Digital Services Act and the Digital Markets Act, with a view to strengthening the Single Market for digital services by creating a safer digital space and a level playing field to foster innovation and competitiveness.
    [Show full text]
  • Engaging Europe: a Transatlantic Digital Agenda for the Biden Administration
    ISSUE BRIEF Engaging Europe: A Transatlantic Digital Agenda for the Biden Administration DECEMBER 2020 FRANCES G. BURWELL INTRODUCTION hile the United States has been embroiled in electoral politics throughout 2020, the European Union (EU) has spent the year laying out an ambitious agenda for regulating digital technology and policy. The next two years will see the EU adopt a raft of Wnew laws with impacts far beyond Europe; indeed, EU leaders have not been shy about their ambitions to create the “gold standard” of digital regulation for the world. The Joe Biden administration must engage with the EU on these issues promptly and energetically, and with a clear understanding of its own ambitions in digital policy. Without that engagement, Europe and the United States are likely to continue their disparate approaches to the digital economy, resulting in ever greater transatlantic tensions over digital tax, privacy, content moderation, competition policy, emerging technologies, and other issues. They The Atlantic Council’s Transatlantic will also lose a vital opportunity to unite in ensuring that authoritarian states— Digital Marketplace Initiative seeks to including China, with its strong tech sector—do not become the rule makers of foster greater US-EU understanding and the global digital world. collaboration on digital policy matters and makes recommendations for building cooperation and ameliorating Digital and tech policy is often treated as a relatively inconsequential sideshow differences in this fast-growing area of in the transatlantic relationship, relegated to discussions among technical the transatlantic economy. experts rather than leadership meetings and summits. As a consequence, even administrations that sought to build a strong and positive partnership with Europe The Future Europe Initiative conducts research and uses real-time commentary have found themselves at odds over digital issues, especially the transfer of data.
    [Show full text]
  • Statement on the Proposals for a Digital Services Act (DSA)
    ERGA STATEMENT ON THE EUROPEAN COMMISSION’S PROPOSALS FOR A ‘DIGITAL SERVICES ACT’ (DSA) AND A ‘DIGITAL MARKETS ACT’ (DMA) Introduction On December 15th 2020, the European Commission released its proposals for a Digital Services Act (DSA) and for a Digital Markets Act (DMA). While these texts are now being reviewed by the European Parliament and the Council respectively, ERGA welcomes the opportunity to express its preliminary views on the proposals. Prior to the release of the legislative proposal, ERGA had published a first Position Paper on the DSA1 and responded to the Commission’s public consultation2. The views expressed were informed by the long-standing experience of ERGA members – as the independent regulatory authorities responsible for audiovisual media services in the EU Member States – in regulating media content and aimed to support the Commission’s reflections on the implications of changes in how online content, whatever its forms, is delivered to European citizens within the Single Market. Fundamental to ERGA’s approach is the idea that the principles underpinning media regulation – including the independence of regulatory authorities - are largely transferable to the online environment. Furthermore, ERGA firmly believes that given its crucial role in promoting and upholding fundamental European values, such as freedom of expression, cultural diversity and human dignity, online content regulation requires tailored rules to address specific challenges. To continue its contribution to public debates on matters relating to online content regulation3, and in line with its advisory role to the European Commission, ERGA sets out below its preliminary assessment of the draft DSA Regulation. The latter is largely based on ERGA members’ expertise in the implementation of legislation supporting content regulation, in particular when it comes to balancing important democratic values with citizens’ fundamental rights.
    [Show full text]
  • Digital Services Act and Digital Markets Act
    Digital Services Act and Digital Markets Act A GUIDE TO THE KEY PROPOSALS & IDEAS FOR IMPROVEMENT Policy Paper 2 December 15, 2020 marked both the end of a long process and the beginning of a new phase of political debate regarding the Internet econ- omy, as European Commissioners Margrethe Vestager and Thierry Breton presented the Digital Services Act (DSA) and the Digital Markets Act (DMA). The proposed regulations will affect everyone on the Internet, from large online services to startups and the average consumer. Strengthening Europe’s digital sovereignty, protecting democracy, enabling competition, harmonizing national regulations, fighting crime, empowering users – with the DSA and DMA, the Commission has set out to address so many issues that one German commentator deemed the plan an “eierlegende Prof. Dr. Woll milch sau,” i an egg-laying pig that also gives wool and milk. It is a Friedbert Pflüger beloved German expression that hints at the impossibility of achieving Chairman all goals with just a single solution. Internet Economy Foundation Yet the scope of the regulations is understandable. We ourselves have detailed in our studies Fair Play in the Digital Arena. How Europe Can Set the Right Framework for Platforms (2016) and Democracy and Digital Disinformation (2020) the various problems that have arisen from large- ly unregulated online services. The 2000 E-commerce Directive, the last major legal framework established by the EU in this context, intro- duced many positive principles regarding liability as well as notice and takedown – which are carried forward in the new DSA proposal. But the E-commerce Directive did not anticipate the rise of large online platforms and its enormous consequences for competition and democracy.
    [Show full text]
  • Free Speech in the Algorithmic Society: Big Data, Private Governance, and New School Speech Regulation
    Free Speech in the Algorithmic Society: Big Data, Private Governance, and New School Speech Regulation Jack M. Balkin* TABLE OF CONTENTS INTRODUCTION ................................................................................. 1151 I. BIG DATA IS SOYLENT GREEN ................................................. 1154 A. Technology Mediates (and Constitutes) Relationships of Power Between People ..................................................... 1157 B. Information Fiduciaries ................................................... 1160 C. Algorithmic Nuisance ...................................................... 1163 D. Personal Robots ............................................................... 1168 II. NEW SCHOOL SPEECH REGULATION AND PRIVATE GOVERNANCE ......................................................................... 1172 A. The Right to Be Forgotten and the Problem of Fake News 1172 B. New School Speech Regulation ......................................... 1173 1. Collateral Censorship ............................................... 1176 2. Digital Prior Restraint............................................... 1177 C. Public Private Cooperation/Cooptation ............................ 1179 III. PRIVATE GOVERNANCE ........................................................... 1182 A. From Game Gods to Social Media .................................... 1184 B. From the Dyadic to the Pluralist Model of Speech Governance ..................................................................... 1186 C. Problems of the New System of
    [Show full text]