Instruction pre-training: Language models are supervised multitask learners

D Cheng, Y Gu, S Huang, J Bi, M Huang… - arxiv preprint arxiv …, 2024 - arxiv.org
Unsupervised multitask pre-training has been the critical method behind the recent success
of language models (LMs). However, supervised multitask learning still holds significant …

A survey on data synthesis and augmentation for large language models

K Wang, J Zhu, M Ren, Z Liu, S Li, Z Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
The success of Large Language Models (LLMs) is inherently linked to the availability of vast,
diverse, and high-quality data for training and evaluation. However, the growth rate of high …

Pedagogical alignment of large language models (llm) for personalized learning: a survey, trends and challenges

MA Razafinirina, WG Dimbisoa, T Mahatody - Journal of Intelligent …, 2024 - scirp.org
This survey paper investigates how personalized learning offered by Large Language
Models (LLMs) could transform educational experiences. We explore Knowledge Editing …

Synthesizrr: Generating diverse datasets with retrieval augmentation

A Divekar, G Durrett - arxiv preprint arxiv:2405.10040, 2024 - arxiv.org
It is often desirable to distill the capabilities of large language models (LLMs) into smaller
student models due to compute and memory constraints. One way to do this for classification …

Advancing large language model attribution through self-improving

L Huang, X Feng, W Ma, L Zhao, Y Fan… - arxiv preprint arxiv …, 2024 - arxiv.org
Teaching large language models (LLMs) to generate text with citations to evidence sources
can mitigate hallucinations and enhance verifiability in information-seeking systems …

Code needs comments: Enhancing code llms with comment augmentation

D Song, H Guo, Y Zhou, S **ng, Y Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
The programming skill is one crucial ability for Large Language Models (LLMs),
necessitating a deep understanding of programming languages (PLs) and their correlation …

Importance weighting can help large language models self-improve

C Jiang, C Chan, W Xue, Q Liu, Y Guo - arxiv preprint arxiv:2408.09849, 2024 - arxiv.org
Large language models (LLMs) have shown remarkable capability in numerous tasks and
applications. However, fine-tuning LLMs using high-quality datasets under external …

LexC-Gen: Generating Data for Extremely Low-Resource Languages with Large Language Models and Bilingual Lexicons

ZX Yong, C Menghini, SH Bach - arxiv preprint arxiv:2402.14086, 2024 - arxiv.org
Data scarcity in low-resource languages can be addressed with word-to-word translations
from labeled task data in high-resource languages using bilingual lexicons. However …

Learning to generate instruction tuning datasets for zero-shot task adaptation

NV Nayak, Y Nan, A Trost, SH Bach - arxiv preprint arxiv:2402.18334, 2024 - arxiv.org
We introduce Bonito, an open-source model for conditional task generation that converts
unannotated text into task-specific training datasets for instruction tuning. We aim to enable …

FewFedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning

Z Zhang, J Zhang, J Huang, L Qu, H Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Instruction tuning has been identified as a crucial technique for optimizing the performance
of large language models (LLMs) in generating human-aligned responses. Nonetheless …