Communication-efficient distributed deep learning: A comprehensive survey
Distributed deep learning (DL) has become prevalent in recent years to reduce training time
by leveraging multiple computing devices (eg, GPUs/TPUs) due to larger models and …
by leveraging multiple computing devices (eg, GPUs/TPUs) due to larger models and …
A guide through the zoo of biased SGD
Abstract Stochastic Gradient Descent (SGD) is arguably the most important single algorithm
in modern machine learning. Although SGD with unbiased gradient estimators has been …
in modern machine learning. Although SGD with unbiased gradient estimators has been …
SoteriaFL: A unified framework for private federated learning with communication compression
To enable large-scale machine learning in bandwidth-hungry environments such as
wireless networks, significant progress has been made recently in designing communication …
wireless networks, significant progress has been made recently in designing communication …
EF21-P and friends: Improved theoretical communication complexity for distributed optimization with bidirectional compression
In this work we focus our attention on distributed optimization problems in the context where
the communication time between the server and the workers is non-negligible. We obtain …
the communication time between the server and the workers is non-negligible. We obtain …
Momentum provably improves error feedback!
Due to the high communication overhead when training machine learning models in a
distributed environment, modern algorithms invariably rely on lossy communication …
distributed environment, modern algorithms invariably rely on lossy communication …
BEER: Fast Rate for Decentralized Nonconvex Optimization with Communication Compression
Communication efficiency has been widely recognized as the bottleneck for large-scale
decentralized machine learning applications in multi-agent or federated environments. To …
decentralized machine learning applications in multi-agent or federated environments. To …
Coresets for Vertical Federated Learning: Regularized Linear Regression and -Means Clustering
Vertical federated learning (VFL), where data features are stored in multiple parties
distributively, is an important area in machine learning. However, the communication …
distributively, is an important area in machine learning. However, the communication …
Lower bounds and nearly optimal algorithms in distributed learning with communication compression
Recent advances in distributed optimization and learning have shown that communication
compression is one of the most effective means of reducing communication. While there …
compression is one of the most effective means of reducing communication. While there …
Communication acceleration of local gradient methods via an accelerated primal-dual algorithm with an inexact prox
Inspired by a recent breakthrough of Mishchenko et al.[2022], who for the first time showed
that local gradient steps can lead to provable communication acceleration, we propose an …
that local gradient steps can lead to provable communication acceleration, we propose an …
Communication-efficient federated learning: A variance-reduced stochastic approach with adaptive sparsification
Federated learning (FL) is an emerging distributed machine learning paradigm that aims to
realize model training without gathering the data from data sources to a central processing …
realize model training without gathering the data from data sources to a central processing …