A survey on in-context learning
With the increasing capabilities of large language models (LLMs), in-context learning (ICL)
has emerged as a new paradigm for natural language processing (NLP), where LLMs make …
has emerged as a new paradigm for natural language processing (NLP), where LLMs make …
The mystery of in-context learning: A comprehensive survey on interpretation and analysis
Understanding in-context learning (ICL) capability that enables large language models
(LLMs) to excel in proficiency through demonstration examples is of utmost importance. This …
(LLMs) to excel in proficiency through demonstration examples is of utmost importance. This …
In-context convergence of transformers
Transformers have recently revolutionized many domains in modern machine learning and
one salient discovery is their remarkable in-context learning capability, where models can …
one salient discovery is their remarkable in-context learning capability, where models can …
Linear Transformers are Versatile In-Context Learners
Recent research has demonstrated that transformers, particularly linear attention models,
implicitly execute gradient-descent-like algorithms on data provided in-context during their …
implicitly execute gradient-descent-like algorithms on data provided in-context during their …
What Makes Multimodal In-Context Learning Work?
Abstract Large Language Models have demonstrated remarkable performance across
various tasks exhibiting the capacity to swiftly acquire new skills such as through In-Context …
various tasks exhibiting the capacity to swiftly acquire new skills such as through In-Context …
Large language models for social networks: Applications, challenges, and solutions
J Zeng, R Huang, W Malik, L Yin, B Babic… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) are transforming the way people generate, explore, and
engage with content. We study how we can develop LLM applications for online social …
engage with content. We study how we can develop LLM applications for online social …
M2qa: Multi-domain multilingual question answering
Generalization and robustness to input variation are core desiderata of machine learning
research. Language varies along several axes, most importantly, language instance (eg …
research. Language varies along several axes, most importantly, language instance (eg …
Competition-level problems are effective llm evaluators
Large language models (LLMs) have demonstrated impressive reasoning capabilities, yet
there is ongoing debate about these abilities and the potential data contamination problem …
there is ongoing debate about these abilities and the potential data contamination problem …
Compositional generative modeling: A single model is not all you need
Large monolithic generative models trained on massive amounts of data have become an
increasingly dominant approach in AI research. In this paper, we argue that we should …
increasingly dominant approach in AI research. In this paper, we argue that we should …
Astraios: Parameter-Efficient Instruction Tuning Code Large Language Models
The high cost of full-parameter fine-tuning (FFT) of Large Language Models (LLMs) has led
to a series of parameter-efficient fine-tuning (PEFT) methods. However, it remains unclear …
to a series of parameter-efficient fine-tuning (PEFT) methods. However, it remains unclear …