Rethinking gradient sparsification as total error minimization

A Sahu, A Dutta, AM Abdelmoniem… - Advances in …, 2021 - proceedings.neurips.cc
Gradient compression is a widely-established remedy to tackle the communication
bottleneck in distributed training of large deep neural networks (DNNs). Under the error …

Hi-speed dnn training with espresso: Unleashing the full potential of gradient compression with near-optimal usage strategies

Z Wang, H Lin, Y Zhu, TSE Ng - Proceedings of the Eighteenth …, 2023 - dl.acm.org
Gradient compression (GC) is a promising approach to addressing the communication
bottleneck in distributed deep learning (DDL). It saves the communication time, but also …

Accelerating model training in multi-cluster environments with consumer-grade gpus

H Lim, J Ye, S Abdu Jyothi, D Han - Proceedings of the ACM SIGCOMM …, 2024 - dl.acm.org
Rapid advances in machine learning necessitate significant computing power and memory
for training, which is accessible only to large corporations today. Small-scale players like …

QUIC-FL: Quick Unbiased Compression for Federated Learning

RB Basat, S Vargaftik, A Portnoy, G Einziger… - arxiv preprint arxiv …, 2022 - arxiv.org
Distributed Mean Estimation (DME), in which $ n $ clients communicate vectors to a
parameter server that estimates their average, is a fundamental building block in …

Communication-compressed adaptive gradient method for distributed nonconvex optimization

Y Wang, L Lin, J Chen - International Conference on Artificial …, 2022 - proceedings.mlr.press
Due to the explosion in the size of the training datasets, distributed learning has received
growing interest in recent years. One of the major bottlenecks is the large communication …

Bytecomp: Revisiting gradient compression in distributed training

Z Wang, H Lin, Y Zhu, TS Ng - arxiv preprint arxiv:2205.14465, 2022 - arxiv.org
Gradient compression (GC) is a promising approach to addressing the communication
bottleneck in distributed deep learning (DDL). However, it is challenging to find the optimal …

Towards Personalized Human Learning at Scale: A Machine Learning Approach

Z Wang - 2023 - search.proquest.com
This thesis focuses on personalized learning in education, a promising and effective means
of learning where the instructions, educational materials, learning paths, analytics, and …

QUIC-FL:: Quick Unbiased Compression for Federated Learning

R Ben-Basat, S Vargaftik, A Portnoy, G Einziger… - 2022 - openreview.net
Distributed Mean Estimation (DME) is a fundamental building block in communication
efficient federated learning. In DME, clients communicate their lossily compressed gradients …

Scaling Deep Learning Through Optimizing Data-and Management-Plane Communications

Z Wang - 2023 - search.proquest.com
Deep neural networks (DNNs) have achieved unparalleled performance in numerous fields,
including computer vision, natural language processing, and recommendation systems …