A survey on in-context learning

Q Dong, L Li, D Dai, C Zheng, J Ma, R Li, H **a… - arxiv preprint arxiv …, 2022 - arxiv.org
With the increasing capabilities of large language models (LLMs), in-context learning (ICL)
has emerged as a new paradigm for natural language processing (NLP), where LLMs make …

The mystery of in-context learning: A comprehensive survey on interpretation and analysis

Y Zhou, J Li, Y **ang, H Yan, L Gui… - Proceedings of the 2024 …, 2024 - aclanthology.org
Understanding in-context learning (ICL) capability that enables large language models
(LLMs) to excel in proficiency through demonstration examples is of utmost importance. This …

In-context convergence of transformers

Y Huang, Y Cheng, Y Liang - arxiv preprint arxiv:2310.05249, 2023 - arxiv.org
Transformers have recently revolutionized many domains in modern machine learning and
one salient discovery is their remarkable in-context learning capability, where models can …

Linear Transformers are Versatile In-Context Learners

M Vladymyrov, J Von Oswald… - Advances in Neural …, 2025 - proceedings.neurips.cc
Recent research has demonstrated that transformers, particularly linear attention models,
implicitly execute gradient-descent-like algorithms on data provided in-context during their …

What Makes Multimodal In-Context Learning Work?

FB Baldassini, M Shukor, M Cord… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Large Language Models have demonstrated remarkable performance across
various tasks exhibiting the capacity to swiftly acquire new skills such as through In-Context …

Large language models for social networks: Applications, challenges, and solutions

J Zeng, R Huang, W Malik, L Yin, B Babic… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) are transforming the way people generate, explore, and
engage with content. We study how we can develop LLM applications for online social …

M2qa: Multi-domain multilingual question answering

L Engländer, H Sterz, C Poth, J Pfeiffer… - arxiv preprint arxiv …, 2024 - arxiv.org
Generalization and robustness to input variation are core desiderata of machine learning
research. Language varies along several axes, most importantly, language instance (eg …

Competition-level problems are effective llm evaluators

Y Huang, Z Lin, X Liu, Y Gong, S Lu, F Lei… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated impressive reasoning capabilities, yet
there is ongoing debate about these abilities and the potential data contamination problem …

Compositional generative modeling: A single model is not all you need

Y Du, L Kaelbling - arxiv preprint arxiv:2402.01103, 2024 - arxiv.org
Large monolithic generative models trained on massive amounts of data have become an
increasingly dominant approach in AI research. In this paper, we argue that we should …

Astraios: Parameter-Efficient Instruction Tuning Code Large Language Models

TY Zhuo, A Zebaze, N Suppattarachai… - arxiv preprint arxiv …, 2024 - arxiv.org
The high cost of full-parameter fine-tuning (FFT) of Large Language Models (LLMs) has led
to a series of parameter-efficient fine-tuning (PEFT) methods. However, it remains unclear …