A comprehensive survey on training acceleration for large machine learning models in IoT

H Wang, Z Qu, Q Zhou, H Zhang, B Luo… - IEEE Internet of …, 2021 - ieeexplore.ieee.org
The ever-growing artificial intelligence (AI) applications have greatly reshaped our world in
many areas, eg, smart home, computer vision, natural language processing, etc. Behind …

Challenges, applications and design aspects of federated learning: A survey

KMJ Rahman, F Ahmed, N Akhter, M Hasan… - IEEe …, 2021 - ieeexplore.ieee.org
Federated learning (FL) is a new technology that has been a hot research topic. It enables
the training of an algorithm across multiple decentralized edge devices or servers holding …

Handling privacy-sensitive medical data with federated learning: challenges and future directions

O Aouedi, A Sacco, K Piamrat… - IEEE journal of …, 2022 - ieeexplore.ieee.org
Recent medical applications are largely dominated by the application of Machine Learning
(ML) models to assist expert decisions, leading to disruptive innovations in radiology …

Draco: Byzantine-resilient distributed training via redundant gradients

L Chen, H Wang, Z Charles… - … on Machine Learning, 2018 - proceedings.mlr.press
Distributed model training is vulnerable to byzantine system failures and adversarial
compute nodes, ie, nodes that use malicious updates to corrupt the global model stored at a …

Slow and stale gradients can win the race: Error-runtime trade-offs in distributed SGD

S Dutta, G Joshi, S Ghosh, P Dube… - International …, 2018 - proceedings.mlr.press
Abstract Distributed Stochastic Gradient Descent (SGD) when run in a synchronous manner,
suffers from delays in waiting for the slowest learners (stragglers). Asynchronous methods …

Gradient coding from cyclic MDS codes and expander graphs

N Raviv, R Tandon, A Dimakis… - … Conference on Machine …, 2018 - proceedings.mlr.press
Gradient coding is a technique for straggler mitigation in distributed learning. In this paper
we design novel gradient codes using tools from classical coding theory, namely, cyclic …

Coded computing for low-latency federated learning over wireless edge networks

S Prakash, S Dhakal, MR Akdeniz… - IEEE Journal on …, 2020 - ieeexplore.ieee.org
Federated learning enables training a global model from data located at the client nodes,
without data sharing and moving client data to a centralized server. Performance of …

Communication-computation efficient gradient coding

M Ye, E Abbe - International Conference on Machine …, 2018 - proceedings.mlr.press
This paper develops coding techniques to reduce the running time of distributed learning
tasks. It characterizes the fundamental tradeoff to compute gradients in terms of three …