Multifacets of lossy compression for scientific data in the Joint-Laboratory of Extreme Scale Computing

F Cappello, M Acosta, E Agullo, H Anzt… - Future Generation …, 2025 - Elsevier
Abstract The Joint Laboratory on Extreme-Scale Computing (JLESC) was initiated at the
same time lossy compression for scientific data became an important topic for the scientific …

Concealing compression-accelerated i/o for hpc applications through in situ task scheduling

S **, S Di, F Vivien, D Wang, Y Robert, D Tao… - Proceedings of the …, 2024 - dl.acm.org
Lossy compression and asynchronous I/O are two of the most effective solutions for reducing
storage overhead and enhancing I/O performance in large-scale high-performance …

Comet: a novel memory-efficient deep learning training framework by using error-bounded lossy compression

S **, C Zhang, X Jiang, Y Feng, H Guan, G Li… - arxiv preprint arxiv …, 2021 - arxiv.org
Training wide and deep neural networks (DNNs) require large amounts of storage resources
such as memory because the intermediate activation data must be saved in the memory …

Ac-gc: Lossy activation compression with guaranteed convergence

RD Evans, T Aamodt - Advances in Neural Information …, 2021 - proceedings.neurips.cc
Parallel hardware devices (eg, graphics processor units) have limited high-bandwidth
memory capacity. This negatively impacts the training of deep neural networks (DNNs) by …

Accelerating parallel write via deeply integrating predictive lossy compression with HDF5

S **, D Tao, H Tang, S Di, S Byna… - … Conference for High …, 2022 - ieeexplore.ieee.org
Lossy compression is one of the most efficient solutions to reduce storage overhead and
improve I/O performance for HPC applications. However, existing parallel I/O libraries …

Fine-tuning language models over slow networks using activation quantization with guarantees

J Wang, B Yuan, L Rimanic, Y He… - Advances in …, 2022 - proceedings.neurips.cc
Communication compression is a crucial technique for modern distributed learning systems
to alleviate their communication bottlenecks over slower networks. Despite recent intensive …

Understanding The Effectiveness of Lossy Compression in Machine Learning Training Sets

R Underwood, JC Calhoun, S Di, F Cappello - arxiv preprint arxiv …, 2024 - arxiv.org
Learning and Artificial Intelligence (ML/AI) techniques have become increasingly prevalent
in high performance computing (HPC). However, these methods depend on vast volumes of …

η-lstm: Co-designing highly-efficient large lstm training via exploiting memory-saving and architectural design opportunities

X Zhang, H **a, D Zhuang, H Sun, X Fu… - 2021 ACM/IEEE 48th …, 2021 - ieeexplore.ieee.org
Recently, the recurrent neural network, or its most popular type—the Long Short Term
Memory (LSTM) network—has achieved great success in a broad spectrum of real-world …

Fine-tuning language models over slow networks using activation compression with guarantees

J Wang, B Yuan, L Rimanic, Y He, T Dao… - arxiv preprint arxiv …, 2022 - arxiv.org
Communication compression is a crucial technique for modern distributed learning systems
to alleviate their communication bottlenecks over slower networks. Despite recent intensive …

Efficient deep neural network training via decreasing precision with layer capacity

A Shen, Z Lai, T Sun, S Li, K Ge, W Liu, D Li - Frontiers of Computer …, 2025 - Springer
Low-precision training has emerged as a practical approach, saving the cost of time,
memory, and energy during deep neural networks (DNNs) training. Typically, the use of …