Communication-efficient distributed deep learning: A comprehensive survey

Z Tang, S Shi, W Wang, B Li, X Chu - arxiv preprint arxiv:2003.06307, 2020 - arxiv.org
Distributed deep learning (DL) has become prevalent in recent years to reduce training time
by leveraging multiple computing devices (eg, GPUs/TPUs) due to larger models and …

A guide through the zoo of biased SGD

Y Demidovich, G Malinovsky… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Stochastic Gradient Descent (SGD) is arguably the most important single algorithm
in modern machine learning. Although SGD with unbiased gradient estimators has been …

SoteriaFL: A unified framework for private federated learning with communication compression

Z Li, H Zhao, B Li, Y Chi - Advances in Neural Information …, 2022 - proceedings.neurips.cc
To enable large-scale machine learning in bandwidth-hungry environments such as
wireless networks, significant progress has been made recently in designing communication …

EF21-P and friends: Improved theoretical communication complexity for distributed optimization with bidirectional compression

K Gruntkowska, A Tyurin… - … Conference on Machine …, 2023 - proceedings.mlr.press
In this work we focus our attention on distributed optimization problems in the context where
the communication time between the server and the workers is non-negligible. We obtain …

Momentum provably improves error feedback!

I Fatkhullin, A Tyurin, P Richtárik - Advances in Neural …, 2024 - proceedings.neurips.cc
Due to the high communication overhead when training machine learning models in a
distributed environment, modern algorithms invariably rely on lossy communication …

BEER: Fast Rate for Decentralized Nonconvex Optimization with Communication Compression

H Zhao, B Li, Z Li, P Richtárik… - Advances in Neural …, 2022 - proceedings.neurips.cc
Communication efficiency has been widely recognized as the bottleneck for large-scale
decentralized machine learning applications in multi-agent or federated environments. To …

Coresets for Vertical Federated Learning: Regularized Linear Regression and -Means Clustering

L Huang, Z Li, J Sun, H Zhao - Advances in Neural …, 2022 - proceedings.neurips.cc
Vertical federated learning (VFL), where data features are stored in multiple parties
distributively, is an important area in machine learning. However, the communication …

Lower bounds and nearly optimal algorithms in distributed learning with communication compression

X Huang, Y Chen, W Yin… - Advances in Neural …, 2022 - proceedings.neurips.cc
Recent advances in distributed optimization and learning have shown that communication
compression is one of the most effective means of reducing communication. While there …

Communication acceleration of local gradient methods via an accelerated primal-dual algorithm with an inexact prox

A Sadiev, D Kovalev… - Advances in Neural …, 2022 - proceedings.neurips.cc
Inspired by a recent breakthrough of Mishchenko et al.[2022], who for the first time showed
that local gradient steps can lead to provable communication acceleration, we propose an …

Communication-efficient federated learning: A variance-reduced stochastic approach with adaptive sparsification

B Wang, J Fang, H Li, B Zeng - IEEE Transactions on Signal …, 2023 - ieeexplore.ieee.org
Federated learning (FL) is an emerging distributed machine learning paradigm that aims to
realize model training without gathering the data from data sources to a central processing …