Multi-Agent Collaborative Data Selection for Efficient LLM Pretraining

T Bai, L Yang, ZH Wong, J Peng, X Zhuang… - arxiv preprint arxiv …, 2024 - arxiv.org
Efficient data selection is crucial to accelerate the pretraining of large language models
(LLMs). While various methods have been proposed to enhance data efficiency, limited …

Stepwise Perplexity-Guided Refinement for Efficient Chain-of-Thought Reasoning in Large Language Models

Y Cui, P He, J Zeng, H Liu, X Tang, Z Dai, Y Han… - arxiv preprint arxiv …, 2025 - arxiv.org
Chain-of-Thought (CoT) reasoning, which breaks down complex tasks into intermediate
reasoning steps, has significantly enhanced the performance of large language models …