Decentralized federated learning: Fundamentals, state of the art, frameworks, trends, and challenges
In recent years, Federated Learning (FL) has gained relevance in training collaborative
models without sharing sensitive data. Since its birth, Centralized FL (CFL) has been the …
models without sharing sensitive data. Since its birth, Centralized FL (CFL) has been the …
A survey of distributed optimization
In distributed optimization of multi-agent systems, agents cooperate to minimize a global
function which is a sum of local objective functions. Motivated by applications including …
function which is a sum of local objective functions. Motivated by applications including …
Federated learning with partial model personalization
We consider two federated learning algorithms for training partially personalized models,
where the shared and personal parameters are updated either simultaneously or alternately …
where the shared and personal parameters are updated either simultaneously or alternately …
An ensemble of differential evolution and Adam for training feed-forward neural networks
Adam is an adaptive gradient descent approach that is commonly used in back-propagation
(BP) algorithms for training feed-forward neural networks (FFNNs). However, it has the …
(BP) algorithms for training feed-forward neural networks (FFNNs). However, it has the …
Decentralized federated averaging
Federated averaging (FedAvg) is a communication-efficient algorithm for distributed training
with an enormous number of clients. In FedAvg, clients keep their data locally for privacy …
with an enormous number of clients. In FedAvg, clients keep their data locally for privacy …
Can decentralized algorithms outperform centralized algorithms? a case study for decentralized parallel stochastic gradient descent
Most distributed machine learning systems nowadays, including TensorFlow and CNTK, are
built in a centralized fashion. One bottleneck of centralized algorithms lies on high …
built in a centralized fashion. One bottleneck of centralized algorithms lies on high …
cpSGD: Communication-efficient and differentially-private distributed SGD
Distributed stochastic gradient descent is an important subroutine in distributed learning. A
setting of particular interest is when the clients are mobile devices, where two important …
setting of particular interest is when the clients are mobile devices, where two important …
Asynchronous decentralized parallel stochastic gradient descent
Most commonly used distributed machine learning systems are either synchronous or
centralized asynchronous. Synchronous algorithms like AllReduce-SGD perform poorly in a …
centralized asynchronous. Synchronous algorithms like AllReduce-SGD perform poorly in a …
Achieving geometric convergence for distributed optimization over time-varying graphs
This paper considers the problem of distributed optimization over time-varying graphs. For
the case of undirected graphs, we introduce a distributed algorithm, referred to as DIGing …
the case of undirected graphs, we introduce a distributed algorithm, referred to as DIGing …
Harnessing smoothness to accelerate distributed optimization
There has been a growing effort in studying the distributed optimization problem over a
network. The objective is to optimize a global function formed by a sum of local functions …
network. The objective is to optimize a global function formed by a sum of local functions …