Large language models for data annotation and synthesis: A survey

Z Tan, D Li, S Wang, A Beigi, B Jiang… - arxiv preprint arxiv …, 2024 - arxiv.org
Data annotation and synthesis generally refers to the labeling or generating of raw data with
relevant information, which could be used for improving the efficacy of machine learning …

Towards a unified view of preference learning for large language models: A survey

B Gao, F Song, Y Miao, Z Cai, Z Yang, L Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) exhibit remarkably powerful capabilities. One of the crucial
factors to achieve success is aligning the LLM's output with human preferences. This …

Bpo: Towards balanced preference optimization between knowledge breadth and depth in alignment

S Wang, Y Tong, H Zhang, D Li, X Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Reinforcement Learning with Human Feedback (RLHF) is the key to the success of large
language models (LLMs) in recent years. In this work, we first introduce the concepts of …

LLM-DER: A Named Entity Recognition Method Based on Large Language Models for Chinese Coal Chemical Domain

L **ao, Y Xu, J Zhao - arxiv preprint arxiv:2409.10077, 2024 - arxiv.org
Domain-specific Named Entity Recognition (NER), whose goal is to recognize domain-
specific entities and their categories, provides an important support for constructing domain …

AIDE: Task-Specific Fine Tuning with Attribute Guided Multi-Hop Data Expansion

J Li, X Zhu, F Liu, Y Qi - arxiv preprint arxiv:2412.06136, 2024 - arxiv.org
Fine-tuning large language models (LLMs) for specific tasks requires high-quality, diverse
training data relevant to the task. Recent research has leveraged LLMs to synthesize …

Annotation-Efficient Language Model Alignment via Diverse and Representative Response Texts

Y **nai, U Honda - openreview.net
Preference optimization is a standard approach to fine-tuning large language models to
align with human preferences. The quantity, diversity, and representativeness of the …