Unleashing the power of data tsunami: A comprehensive survey on data assessment and selection for instruction tuning of language models
Instruction tuning plays a critical role in aligning large language models (LLMs) with human
preference. Despite the vast amount of open instruction datasets, naively training a LLM on …
preference. Despite the vast amount of open instruction datasets, naively training a LLM on …
Self-exploring language models: Active preference elicitation for online alignment
Preference optimization, particularly through Reinforcement Learning from Human
Feedback (RLHF), has achieved significant success in aligning Large Language Models …
Feedback (RLHF), has achieved significant success in aligning Large Language Models …
Data management for large language models: A survey
Data plays a fundamental role in the training of Large Language Models (LLMs). Effective
data management, particularly in the formulation of a well-suited training dataset, holds …
data management, particularly in the formulation of a well-suited training dataset, holds …
Causal prompting: Debiasing large language model prompting based on front-door adjustment
Despite the notable advancements of existing prompting methods, such as In-Context
Learning and Chain-of-Thought for Large Language Models (LLMs), they still face …
Learning and Chain-of-Thought for Large Language Models (LLMs), they still face …
LLMs-as-Instructors: learning from errors toward automating model improvement
This paper introduces the innovative" LLMs-as-Instructors" framework, which leverages the
advanced Large Language Models (LLMs) to autonomously enhance the training of smaller …
advanced Large Language Models (LLMs) to autonomously enhance the training of smaller …
Knowagent: Knowledge-augmented planning for llm-based agents
Large Language Models (LLMs) have demonstrated great potential in complex reasoning
tasks, yet they fall short when tackling more sophisticated challenges, especially when …
tasks, yet they fall short when tackling more sophisticated challenges, especially when …
Rethinking data selection at scale: Random selection is almost all you need
Supervised fine-tuning (SFT) is crucial for aligning Large Language Models (LLMs) with
human instructions. The primary goal during SFT is to select a small yet representative …
human instructions. The primary goal during SFT is to select a small yet representative …
Federated Data-Efficient Instruction Tuning for Large Language Models
Instruction tuning helps improve pretrained large language models (LLMs) in terms of the
responsiveness to human instructions, which is benefited from diversified instruction data …
responsiveness to human instructions, which is benefited from diversified instruction data …
SelectIT: Selective Instruction Tuning for Large Language Models via Uncertainty-Aware Self-Reflection
Instruction tuning (IT) is crucial to tailoring large language models (LLMs) towards human-
centric interactions. Recent advancements have shown that the careful selection of a small …
centric interactions. Recent advancements have shown that the careful selection of a small …
Instruction Embedding: Latent Representations of Instructions Towards Task Identification
Instruction data is crucial for improving the capability of Large Language Models (LLMs) to
align with human-level performance. Recent research LIMA demonstrates that alignment is …
align with human-level performance. Recent research LIMA demonstrates that alignment is …