Advances and Open Problems in Federated Learning
Total Page:16
File Type:pdf, Size:1020Kb
Advances and Open Problems in Federated Learning Peter Kairouz7* H. Brendan McMahan7∗ Brendan Avent21 Aurelien´ Bellet9 Mehdi Bennis19 Arjun Nitin Bhagoji13 Kallista Bonawitz7 Zachary Charles7 Graham Cormode23 Rachel Cummings6 Rafael G.L. D’Oliveira14 Hubert Eichner7 Salim El Rouayheb14 David Evans22 Josh Gardner24 Zachary Garrett7 Adria` Gascon´ 7 Badih Ghazi7 Phillip B. Gibbons2 Marco Gruteser7;14 Zaid Harchaoui24 Chaoyang He21 Lie He 4 Zhouyuan Huo 20 Ben Hutchinson7 Justin Hsu25 Martin Jaggi4 Tara Javidi17 Gauri Joshi2 Mikhail Khodak2 Jakub Konecnˇ y´7 Aleksandra Korolova21 Farinaz Koushanfar17 Sanmi Koyejo7;18 Tancrede` Lepoint7 Yang Liu12 Prateek Mittal13 Mehryar Mohri7 Richard Nock1 Ayfer Ozg¨ ur¨ 15 Rasmus Pagh7;10 Hang Qi7 Daniel Ramage7 Ramesh Raskar11 Mariana Raykova7 Dawn Song16 Weikang Song7 Sebastian U. Stich4 Ziteng Sun3 Ananda Theertha Suresh7 Florian Tramer` 15 Praneeth Vepakomma11 Jianyu Wang2 Li Xiong5 Zheng Xu7 Qiang Yang8 Felix X. Yu7 Han Yu12 Sen Zhao7 1Australian National University, 2Carnegie Mellon University, 3Cornell University, 4Ecole´ Polytechnique Fed´ erale´ de Lausanne, 5Emory University, 6Georgia Institute of Technology, 7Google Research, 8Hong Kong University of Science and Technology, 9INRIA, 10IT University of Copenhagen, 11Massachusetts Institute of Technology, 12Nanyang Technological University, 13Princeton University, 14Rutgers University, 15Stanford University, 16University of California Berkeley, 17 University of California San Diego, 18University of Illinois Urbana-Champaign, 19University of Oulu, 20University of Pittsburgh, 21University of Southern California, 22University of Virginia, 23University of Warwick, 24University of Washington, 25University of Wisconsin–Madison arXiv:1912.04977v3 [cs.LG] 9 Mar 2021 Abstract Federated learning (FL) is a machine learning setting where many clients (e.g. mobile devices or whole organizations) collaboratively train a model under the orchestration of a central server (e.g. service provider), while keeping the training data decentralized. FL embodies the principles of focused data collection and minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, centralized machine learning and data science approaches. Motivated by the explosive growth in FL research, this paper discusses recent advances and presents an extensive collection of open problems and challenges. *Peter Kairouz and H. Brendan McMahan conceived, coordinated, and edited this work. Correspondence to kairouz@ google.com and [email protected]. 1 Contents 1 Introduction 4 1.1 The Cross-Device Federated Learning Setting . .5 1.1.1 The Lifecycle of a Model in Federated Learning . .7 1.1.2 A Typical Federated Training Process . .8 1.2 Federated Learning Research . .9 1.3 Organization . 10 2 Relaxing the Core FL Assumptions: Applications to Emerging Settings and Scenarios 11 2.1 Fully Decentralized / Peer-to-Peer Distributed Learning . 11 2.1.1 Algorithmic Challenges . 12 2.1.2 Practical Challenges . 14 2.2 Cross-Silo Federated Learning . 14 2.3 Split Learning . 16 2.4 Executive summary . 17 3 Improving Efficiency and Effectiveness 18 3.1 Non-IID Data in Federated Learning . 18 3.1.1 Strategies for Dealing with Non-IID Data . 19 3.2 Optimization Algorithms for Federated Learning . 20 3.2.1 Optimization Algorithms and Convergence Rates for IID Datasets . 21 3.2.2 Optimization Algorithms and Convergence Rates for Non-IID Datasets . 25 3.3 Multi-Task Learning, Personalization, and Meta-Learning . 28 3.3.1 Personalization via Featurization . 28 3.3.2 Multi-Task Learning . 28 3.3.3 Local Fine Tuning and Meta-Learning . 29 3.3.4 When is a Global FL-trained Model Better? . 30 3.4 Adapting ML Workflows for Federated Learning . 30 3.4.1 Hyperparameter Tuning . 31 3.4.2 Neural Architecture Design . 31 3.4.3 Debugging and Interpretability for FL . 32 3.5 Communication and Compression . 32 3.6 Application To More Types of Machine Learning Problems and Models . 34 3.7 Executive summary . 34 4 Preserving the Privacy of User Data 36 4.1 Actors, Threat Models, and Privacy in Depth . 37 4.2 Tools and Technologies . 38 4.2.1 Secure Computations . 40 4.2.2 Privacy-Preserving Disclosures . 44 4.2.3 Verifiability . 46 4.3 Protections Against External Malicious Actors . 48 4.3.1 Auditing the Iterates and Final Model . 49 4.3.2 Training with Central Differential Privacy . 49 4.3.3 Concealing the Iterates . 51 4.3.4 Repeated Analyses over Evolving Data . 52 4.3.5 Preventing Model Theft and Misuse . 52 4.4 Protections Against an Adversarial Server . 53 4.4.1 Challenges: Communication Channels, Sybil Attacks, and Selection . 53 4.4.2 Limitations of Existing Solutions . 54 4.4.3 Training with Distributed Differential Privacy . 55 4.4.4 Preserving Privacy While Training Sub-Models . 58 2 4.5 User Perception . 59 4.5.1 Understanding Privacy Needs for Particular Analysis Tasks . 59 4.5.2 Behavioral Research to Elicit Privacy Preferences . 60 4.6 Executive Summary . 60 5 Defending Against Attacks and Failures 62 5.1 Adversarial Attacks on Model Performance . 62 5.1.1 Goals and Capabilities of an Adversary . 63 5.1.2 Model Update Poisoning . 66 5.1.3 Data Poisoning Attacks . 67 5.1.4 Inference-Time Evasion Attacks . 69 5.1.5 Defensive Capabilities from Privacy Guarantees . 70 5.2 Non-Malicious Failure Modes . 71 5.3 Exploring the Tension between Privacy and Robustness . 73 5.4 Executive Summary . 73 6 Ensuring Fairness and Addressing Sources of Bias 75 6.1 Bias in Training Data . 75 6.2 Fairness Without Access to Sensitive Attributes . 76 6.3 Fairness, Privacy, and Robustness . 77 6.4 Leveraging Federation to Improve Model Diversity . 78 6.5 Federated Fairness: New Opportunities and Challenges . 79 6.6 Executive Summary . 79 7 Addressing System Challenges 81 7.1 Platform Development and Deployment Challenges . 81 7.2 System Induced Bias . 82 7.2.1 Device Availability Profiles . 82 7.2.2 Examples of System Induced Bias . 83 7.2.3 Open Challenges in Quantifying and Mitigating System Induced Bias . 84 7.3 System Parameter Tuning . 85 7.4 On-Device Runtime . 86 7.5 The Cross-Silo Setting . 87 7.6 Executive Summary . 88 8 Concluding Remarks 89 A Software and Datasets for Federated Learning 119 3 1 Introduction Federated learning (FL) is a machine learning setting where many clients (e.g. mobile devices or whole or- ganizations) collaboratively train a model under the orchestration of a central server (e.g. service provider), while keeping the training data decentralized. It embodies the principles of focused collection and data minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, cen- tralized machine learning. This area has received significant interest recently, both from research and applied perspectives. This paper describes the defining characteristics and challenges of the federated learning set- ting, highlights important practical constraints and considerations, and then enumerates a range of valuable research directions. The goals of this work are to highlight research problems that are of significant theo- retical and practical interest, and to encourage research on problems that could have significant real-world impact. The term federated learning was introduced in 2016 by McMahan et al. [337]: “We term our approach Federated Learning, since the learning task is solved by a loose federation of participating devices (which we refer to as clients) which are coordinated by a central server.” An unbalanced and non-IID (identically and independently distributed) data partitioning across a massive number of unreliable devices with limited communication bandwidth was introduced as the defining set of challenges. Significant related work predates the introduction of the term federated learning. A longstanding goal pursued by many research communities (including cryptography, databases, and machine learning) is to ana- lyze and learn from data distributed among many owners without exposing that data. Cryptographic methods for computing on encrypted data were developed starting in the early 1980s [396, 492], and Agrawal and Srikant [11] and Vaidya et al. [457] are early examples of work that sought to learn from local data using a centralized server while preserving privacy. Conversely, even since the introduction of the term federated learning, we are aware of no single work that directly addresses the full set of FL challenges. Thus, the term federated learning provides a convenient shorthand for a set of characteristics, constraints, and challenges that often co-occur in applied ML problems on decentralized data where privacy is paramount. This paper originated at the Workshop on Federated Learning and Analytics held June 17–18th, 2019, hosted at Google’s Seattle office. During the course of this.