Unleashing the power of data tsunami: A comprehensive survey on data assessment and selection for instruction tuning of language models

Y Qin, Y Yang, P Guo, G Li, H Shao, Y Shi, Z Xu… - arxiv preprint arxiv …, 2024 - arxiv.org
Instruction tuning plays a critical role in aligning large language models (LLMs) with human
preference. Despite the vast amount of open instruction datasets, naively training a LLM on …

Self-exploring language models: Active preference elicitation for online alignment

S Zhang, D Yu, H Sharma, H Zhong, Z Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
Preference optimization, particularly through Reinforcement Learning from Human
Feedback (RLHF), has achieved significant success in aligning Large Language Models …

Data management for large language models: A survey

Z Wang, W Zhong, Y Wang, Q Zhu, F Mi, B Wang… - CoRR, 2023 - openreview.net
Data plays a fundamental role in the training of Large Language Models (LLMs). Effective
data management, particularly in the formulation of a well-suited training dataset, holds …

Causal prompting: Debiasing large language model prompting based on front-door adjustment

C Zhang, L Zhang, J Wu, Y He, D Zhou - arxiv preprint arxiv:2403.02738, 2024 - arxiv.org
Despite the notable advancements of existing prompting methods, such as In-Context
Learning and Chain-of-Thought for Large Language Models (LLMs), they still face …

LLMs-as-Instructors: learning from errors toward automating model improvement

J Ying, M Lin, Y Cao, W Tang, B Wang, Q Sun… - arxiv preprint arxiv …, 2024 - arxiv.org
This paper introduces the innovative" LLMs-as-Instructors" framework, which leverages the
advanced Large Language Models (LLMs) to autonomously enhance the training of smaller …

Knowagent: Knowledge-augmented planning for llm-based agents

Y Zhu, S Qiao, Y Ou, S Deng, N Zhang, S Lyu… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have demonstrated great potential in complex reasoning
tasks, yet they fall short when tackling more sophisticated challenges, especially when …

Rethinking data selection at scale: Random selection is almost all you need

T **a, B Yu, K Dang, A Yang, Y Wu, Y Tian… - arxiv preprint arxiv …, 2024 - arxiv.org
Supervised fine-tuning (SFT) is crucial for aligning Large Language Models (LLMs) with
human instructions. The primary goal during SFT is to select a small yet representative …

Federated Data-Efficient Instruction Tuning for Large Language Models

Z Qin, Z Wu, B He, S Deng - arxiv preprint arxiv:2410.10926, 2024 - arxiv.org
Instruction tuning helps improve pretrained large language models (LLMs) in terms of the
responsiveness to human instructions, which is benefited from diversified instruction data …

SelectIT: Selective Instruction Tuning for Large Language Models via Uncertainty-Aware Self-Reflection

L Liu, X Liu, DF Wong, D Li, Z Wang, B Hu… - arxiv preprint arxiv …, 2024 - arxiv.org
Instruction tuning (IT) is crucial to tailoring large language models (LLMs) towards human-
centric interactions. Recent advancements have shown that the careful selection of a small …

Instruction Embedding: Latent Representations of Instructions Towards Task Identification

Y Li, J Shi, S Feng, P Yuan, X Wang, B Pan… - arxiv preprint arxiv …, 2024 - arxiv.org
Instruction data is crucial for improving the capability of Large Language Models (LLMs) to
align with human-level performance. Recent research LIMA demonstrates that alignment is …