A survey of distributed optimization

T Yang, X Yi, J Wu, Y Yuan, D Wu, Z Meng… - Annual Reviews in …, 2019 - Elsevier
In distributed optimization of multi-agent systems, agents cooperate to minimize a global
function which is a sum of local objective functions. Motivated by applications including …

Distributed optimization for control

A Nedić, J Liu - Annual Review of Control, Robotics, and …, 2018 - annualreviews.org
Advances in wired and wireless technology have necessitated the development of theory,
models, and tools to cope with the new challenges posed by large-scale control and …

Can decentralized algorithms outperform centralized algorithms? a case study for decentralized parallel stochastic gradient descent

X Lian, C Zhang, H Zhang, CJ Hsieh… - Advances in neural …, 2017 - proceedings.neurips.cc
Most distributed machine learning systems nowadays, including TensorFlow and CNTK, are
built in a centralized fashion. One bottleneck of centralized algorithms lies on high …

Network topology and communication-computation tradeoffs in decentralized optimization

A Nedić, A Olshevsky, MG Rabbat - Proceedings of the IEEE, 2018 - ieeexplore.ieee.org
In decentralized optimization, nodes cooperate to minimize an overall objective function that
is the sum (or average) of per-node private objective functions. Algorithms interleave local …

Asynchronous decentralized parallel stochastic gradient descent

X Lian, W Zhang, C Zhang, J Liu - … Conference on Machine …, 2018 - proceedings.mlr.press
Most commonly used distributed machine learning systems are either synchronous or
centralized asynchronous. Synchronous algorithms like AllReduce-SGD perform poorly in a …

Distributed stochastic gradient tracking methods

S Pu, A Nedić - Mathematical Programming, 2021 - Springer
In this paper, we study the problem of distributed multi-agent optimization over a network,
where each agent possesses a local cost function that is smooth and strongly convex. The …

[BOK][B] First-order and stochastic optimization methods for machine learning

G Lan - 2020 - Springer
Since its beginning, optimization has played a vital role in data science. The analysis and
solution methods for many statistical and machine learning models rely on optimization. The …

Robust low-rank tensor recovery with rectification and alignment

X Zhang, D Wang, Z Zhou, Y Ma - IEEE Transactions on Pattern …, 2019 - ieeexplore.ieee.org
Low-rank tensor recovery in the presence of sparse but arbitrary errors is an important
problem with many practical applications. In this work, we propose a general framework that …

Communication-efficient algorithms for decentralized and stochastic optimization

G Lan, S Lee, Y Zhou - Mathematical Programming, 2020 - Springer
We present a new class of decentralized first-order methods for nonsmooth and stochastic
optimization problems defined over multiagent networks. Considering that communication is …

Communication-efficient distributed deep learning: A comprehensive survey

Z Tang, S Shi, W Wang, B Li, X Chu - arxiv preprint arxiv:2003.06307, 2020 - arxiv.org
Distributed deep learning (DL) has become prevalent in recent years to reduce training time
by leveraging multiple computing devices (eg, GPUs/TPUs) due to larger models and …