Federated learning on non-IID data: A survey

H Zhu, J Xu, S Liu, Y ** - Neurocomputing, 2021 - Elsevier
Federated learning is an emerging distributed machine learning framework for privacy
preservation. However, models trained in federated learning usually have worse …

Hardware acceleration of sparse and irregular tensor computations of ml models: A survey and insights

S Dave, R Baghdadi, T Nowatzki… - Proceedings of the …, 2021 - ieeexplore.ieee.org
Machine learning (ML) models are widely used in many important domains. For efficiently
processing these computational-and memory-intensive applications, tensors of these …

Rethinking gradient sparsification as total error minimization

A Sahu, A Dutta, AM Abdelmoniem… - Advances in …, 2021 - proceedings.neurips.cc
Gradient compression is a widely-established remedy to tackle the communication
bottleneck in distributed training of large deep neural networks (DNNs). Under the error …

Time-correlated sparsification for communication-efficient federated learning

E Ozfatura, K Ozfatura, D Gündüz - 2021 IEEE International …, 2021 - ieeexplore.ieee.org
Federated learning (FL) enables multiple clients to collaboratively train a shared model, with
the help of a parameter server (PS), without disclosing their local datasets. However, due to …

Neural gradients are near-lognormal: improved quantized and sparse training

B Chmiel, L Ben-Uri, M Shkolnik, E Hoffer… - arxiv preprint arxiv …, 2020 - arxiv.org
While training can mostly be accelerated by reducing the time needed to propagate neural
gradients back throughout the model, most previous works focus on the quantization/pruning …

Sketch-fusion: a gradient compression method with multi-layer fusion for communication-efficient distributed training

L Dai, L Gong, Z An, Y Xu, B Diao - Journal of Parallel and Distributed …, 2024 - Elsevier
Gradient compression is an effective technique for improving the efficiency of distributed
training. However, introducing gradient compression can reduce model accuracy and …

Fedlite: A scalable approach for federated learning on resource-constrained clients

J Wang, H Qi, AS Rawat, S Reddi, S Waghmare… - arxiv preprint arxiv …, 2022 - arxiv.org
In classical federated learning, the clients contribute to the overall training by communicating
local updates for the underlying model on their private data to a coordinating server …

Distributed artificial intelligence: review, taxonomy, framework, and reference architecture

N Janbi, I Katib, R Mehmood - Taxonomy, Framework, and …, 2023 - papers.ssrn.com
Artificial intelligence (AI) research and market have grown rapidly in the last few years and
this trend is expected to continue with many potential advancements and innovations in this …

Compressed communication for distributed training: Adaptive methods and system

Y Zhong, C **e, S Zheng, H Lin - arxiv preprint arxiv:2105.07829, 2021 - arxiv.org
Communication overhead severely hinders the scalability of distributed machine learning
systems. Recently, there has been a growing interest in using gradient compression to …

Alternate model growth and pruning for efficient training of recommendation systems

X Du, B Bhushanam, J Yu, D Choudhary… - 2021 20th IEEE …, 2021 - ieeexplore.ieee.org
Deep learning recommendation systems at scale have provided remarkable gains through
increasing model capacity (ie wider and deeper neural networks), but it comes at significant …