CAFE: Towards Compact, Adaptive, and Fast Embedding for Large-scale Recommendation Models

H Zhang, Z Liu, B Chen, Y Zhao, T Zhao… - Proceedings of the …, 2024 - dl.acm.org
Recently, the growing memory demands of embedding tables in Deep Learning
Recommendation Models (DLRMs) pose great challenges for model training and …

CAFE+: Towards Compact, Adaptive, and Fast Embedding for Large-scale Online Recommendation Models

Z Liu, H Zhang, B Chen, Z Jiang, Y Zhao… - ACM Transactions on …, 2025 - dl.acm.org
The growing memory demands of embedding tables in Deep Learning Recommendation
Models (DLRMs) pose great challenges for model training and deployment. Existing …

Revisiting Data Analysis with Pre-trained Foundation Models

C Liang, D Yang, Z Liang, Z Liang, T Zhang… - arxiv preprint arxiv …, 2025 - arxiv.org
Data analysis focuses on harnessing advanced statistics, programming, and machine
learning techniques to extract valuable insights from vast datasets. An increasing volume …

From Sancus to Sancus: staleness and quantization-aware full-graph decentralized training in graph neural networks

J Peng, Q Liu, Z Chen, Y Shao, Y Shen, L Chen… - The VLDB Journal, 2025 - Springer
Graph neural networks (GNNs) have emerged due to their success at modeling graph data.
Yet, it is challenging for GNNs to efficiently scale to large graphs. Thus, distributed GNNs …

Surge Phenomenon in Optimal Learning Rate and Batch Size Scaling

S Li, P Zhao, H Zhang, X Sun, H Wu, D Jiao… - arxiv preprint arxiv …, 2024 - arxiv.org
In current deep learning tasks, Adam style optimizers such as Adam, Adagrad, RMSProp,
Adafactor, and Lion have been widely used as alternatives to SGD style optimizers. These …

Zero-Indexing Internet Search Augmented Generation for Large Language Models

G He, Z Dai, J Zhu, B Zhao, C Li, Y Peng… - arxiv preprint arxiv …, 2024 - arxiv.org
Retrieval augmented generation has emerged as an effective method to enhance large
language model performance. This approach typically relies on an internal retrieval module …