Communication-Efficient Federated Learning with Sketching Ashwinee Panda Daniel Rothchild Enayat Ullah Nikita Ivkin Ion Stoica Joseph Gonzalez, Ed. Raman Arora, Ed. Vladimir Braverman, Ed. Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2020-57 http://www2.eecs.berkeley.edu/Pubs/TechRpts/2020/EECS-2020-57.html May 23, 2020 Copyright © 2020, by the author(s). All rights reserved. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission. COMMUNICATION-EFFICIENT FEDERATED LEARNING WITH SKETCHING APREPRINT 1 1 2 3† Daniel Rothchild ∗ Ashwinee Panda ∗ Enayat Ullah Nikita Ivkin Vladimir Braverman2 Joseph Gonzalez1 Ion Stoica1 Raman Arora2 1 UC Berkeley, drothchild, ashwineepanda, jegonzal, istoica @berkeley.edu 2Johns Hopkinsf University,
[email protected], vova, arora @cs.jhu.edug 3 Amazon,
[email protected] g ABSTRACT Existing approaches to federated learning suffer from a communication bottleneck as well as conver- gence issues due to sparse client participation. In this paper we introduce a novel algorithm, called FedSketchedSGD, to overcome these challenges. FedSketchedSGD compresses model updates using a Count Sketch, and then takes advantage of the mergeability of sketches to combine model updates from many workers. A key insight in the design of FedSketchedSGD is that, because the Count Sketch is linear, momentum and error accumulation can both be carried out within the sketch.