A survey on data selection for language models

A Albalak, Y Elazar, SM **e, S Longpre… - arxiv preprint arxiv …, 2024 - arxiv.org
A major factor in the recent success of large language models is the use of enormous and
ever-growing text datasets for unsupervised pre-training. However, naively training a model …

A survey on stability of learning with limited labelled data and its sensitivity to the effects of randomness

B Pecher, I Srba, M Bielikova - ACM Computing Surveys, 2024 - dl.acm.org
Learning with limited labelled data, such as prompting, in-context learning, fine-tuning, meta-
learning, or few-shot learning, aims to effectively train a model using only a small amount of …

A survey on rag meeting llms: Towards retrieval-augmented large language models

W Fan, Y Ding, L Ning, S Wang, H Li, D Yin… - Proceedings of the 30th …, 2024 - dl.acm.org
As one of the most advanced techniques in AI, Retrieval-Augmented Generation (RAG) can
offer reliable and up-to-date external knowledge, providing huge convenience for numerous …

A systematic survey of prompt engineering on vision-language foundation models

J Gu, Z Han, S Chen, A Beirami, B He, G Zhang… - arxiv preprint arxiv …, 2023 - arxiv.org
Prompt engineering is a technique that involves augmenting a large pre-trained model with
task-specific hints, known as prompts, to adapt the model to new tasks. Prompts can be …

Grammar prompting for domain-specific language generation with large language models

B Wang, Z Wang, X Wang, Y Cao… - Advances in Neural …, 2024 - proceedings.neurips.cc
Large language models (LLMs) can learn to perform a wide range of natural language tasks
from just a handful of in-context examples. However, for generating strings from highly …

Label words are anchors: An information flow perspective for understanding in-context learning

L Wang, L Li, D Dai, D Chen, H Zhou, F Meng… - arxiv preprint arxiv …, 2023 - arxiv.org
In-context learning (ICL) emerges as a promising capability of large language models
(LLMs) by providing them with demonstration examples to perform diverse tasks. However …

Learning to retrieve in-context examples for large language models

L Wang, N Yang, F Wei - arxiv preprint arxiv:2307.07164, 2023 - arxiv.org
Large language models (LLMs) have demonstrated their ability to learn in-context, allowing
them to perform various tasks based on a few input-output examples. However, the …

Boosting language models reasoning with chain-of-knowledge prompting

J Wang, Q Sun, X Li, M Gao - arxiv preprint arxiv:2306.06427, 2023 - arxiv.org
Recently, Chain-of-Thought (CoT) prompting has delivered success on complex reasoning
tasks, which aims at designing a simple prompt like``Let's think step by step''or multiple in …

Small models are valuable plug-ins for large language models

C Xu, Y Xu, S Wang, Y Liu, C Zhu… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) such as GPT-3 and GPT-4 are powerful but their weights are
often publicly unavailable and their immense sizes make the models difficult to be tuned with …

Uprise: Universal prompt retrieval for improving zero-shot evaluation

D Cheng, S Huang, J Bi, Y Zhan, J Liu, Y Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) are popular for their impressive abilities, but the need for
model-specific fine-tuning or task-specific prompt engineering can hinder their …