A survey on data selection for language models
A major factor in the recent success of large language models is the use of enormous and
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
A survey on stability of learning with limited labelled data and its sensitivity to the effects of randomness
Learning with limited labelled data, such as prompting, in-context learning, fine-tuning, meta-
learning, or few-shot learning, aims to effectively train a model using only a small amount of …
learning, or few-shot learning, aims to effectively train a model using only a small amount of …
A survey on rag meeting llms: Towards retrieval-augmented large language models
As one of the most advanced techniques in AI, Retrieval-Augmented Generation (RAG) can
offer reliable and up-to-date external knowledge, providing huge convenience for numerous …
offer reliable and up-to-date external knowledge, providing huge convenience for numerous …
A systematic survey of prompt engineering on vision-language foundation models
Prompt engineering is a technique that involves augmenting a large pre-trained model with
task-specific hints, known as prompts, to adapt the model to new tasks. Prompts can be …
task-specific hints, known as prompts, to adapt the model to new tasks. Prompts can be …
Grammar prompting for domain-specific language generation with large language models
Large language models (LLMs) can learn to perform a wide range of natural language tasks
from just a handful of in-context examples. However, for generating strings from highly …
from just a handful of in-context examples. However, for generating strings from highly …
Label words are anchors: An information flow perspective for understanding in-context learning
In-context learning (ICL) emerges as a promising capability of large language models
(LLMs) by providing them with demonstration examples to perform diverse tasks. However …
(LLMs) by providing them with demonstration examples to perform diverse tasks. However …
Learning to retrieve in-context examples for large language models
Large language models (LLMs) have demonstrated their ability to learn in-context, allowing
them to perform various tasks based on a few input-output examples. However, the …
them to perform various tasks based on a few input-output examples. However, the …
Boosting language models reasoning with chain-of-knowledge prompting
Recently, Chain-of-Thought (CoT) prompting has delivered success on complex reasoning
tasks, which aims at designing a simple prompt like``Let's think step by step''or multiple in …
tasks, which aims at designing a simple prompt like``Let's think step by step''or multiple in …
Small models are valuable plug-ins for large language models
Large language models (LLMs) such as GPT-3 and GPT-4 are powerful but their weights are
often publicly unavailable and their immense sizes make the models difficult to be tuned with …
often publicly unavailable and their immense sizes make the models difficult to be tuned with …
Uprise: Universal prompt retrieval for improving zero-shot evaluation
Large Language Models (LLMs) are popular for their impressive abilities, but the need for
model-specific fine-tuning or task-specific prompt engineering can hinder their …
model-specific fine-tuning or task-specific prompt engineering can hinder their …