In-context retrieval-augmented language models

O Ram, Y Levine, I Dalmedigos, D Muhlgay… - Transactions of the …, 2023 - direct.mit.edu
Abstract Retrieval-Augmented Language Modeling (RALM) methods, which condition a
language model (LM) on relevant documents from a grounding corpus during generation …

Holistic evaluation of language models

P Liang, R Bommasani, T Lee, D Tsipras… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models (LMs) are becoming the foundation for almost all major language
technologies, but their capabilities, limitations, and risks are not well understood. We present …

Large language models for information retrieval: A survey

Y Zhu, H Yuan, S Wang, J Liu, W Liu, C Deng… - arxiv preprint arxiv …, 2023 - arxiv.org
As a primary means of information acquisition, information retrieval (IR) systems, such as
search engines, have integrated themselves into our daily lives. These systems also serve …

Large language models are effective text rankers with pairwise ranking prompting

Z Qin, R Jagerman, K Hui, H Zhuang, J Wu… - arxiv preprint arxiv …, 2023 - arxiv.org
Ranking documents using Large Language Models (LLMs) by directly feeding the query and
candidate documents into the prompt is an interesting and practical problem. However …

Autoregressive search engines: Generating substrings as document identifiers

M Bevilacqua, G Ottaviano, P Lewis… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Knowledge-intensive language tasks require NLP systems to both provide the
correct answer and retrieve supporting evidence for it in a given corpus. Autoregressive …

Adapting language models to compress contexts

A Chevalier, A Wettig, A Ajith, D Chen - arxiv preprint arxiv:2305.14788, 2023 - arxiv.org
Transformer-based language models (LMs) are powerful and widely-applicable tools, but
their usefulness is constrained by a finite context window and the expensive computational …

Coder reviewer reranking for code generation

T Zhang, T Yu, T Hashimoto, M Lewis… - International …, 2023 - proceedings.mlr.press
Sampling diverse programs from a code language model and reranking with model
likelihood is a popular method for code generation but it is prone to preferring degenerate …

Rankt5: Fine-tuning t5 for text ranking with ranking losses

H Zhuang, Z Qin, R Jagerman, K Hui, J Ma… - Proceedings of the 46th …, 2023 - dl.acm.org
Pretrained language models such as BERT have been shown to be exceptionally effective
for text ranking. However, there are limited studies on how to leverage more powerful …

A setwise approach for effective and highly efficient zero-shot ranking with large language models

S Zhuang, H Zhuang, B Koopman… - Proceedings of the 47th …, 2024 - dl.acm.org
We propose a novel zero-shot document ranking approach based on Large Language
Models (LLMs): the Setwise prompting approach. Our approach complements existing …

A critical evaluation of evaluations for long-form question answering

F Xu, Y Song, M Iyyer, E Choi - arxiv preprint arxiv:2305.18201, 2023 - arxiv.org
Long-form question answering (LFQA) enables answering a wide range of questions, but its
flexibility poses enormous challenges for evaluation. We perform the first targeted study of …