Information retrieval: recent advances and beyond

KA Hambarde, H Proenca - IEEE Access, 2023 - ieeexplore.ieee.org
This paper provides an extensive and thorough overview of the models and techniques
utilized in the first and second stages of the typical information retrieval processing chain …

Colbertv2: Effective and efficient retrieval via lightweight late interaction

K Santhanam, O Khattab, J Saad-Falcon… - arxiv preprint arxiv …, 2021 - arxiv.org
Neural information retrieval (IR) has greatly advanced search and other knowledge-
intensive language tasks. While many neural IR methods encode queries and documents …

Unsupervised corpus aware language model pre-training for dense passage retrieval

L Gao, J Callan - arxiv preprint arxiv:2108.05540, 2021 - arxiv.org
Recent research demonstrates the effectiveness of using fine-tuned language models~(LM)
for dense retrieval. However, dense retrievers are hard to train, typically requiring heavily …

RocketQAv2: A joint training method for dense passage retrieval and passage re-ranking

R Ren, Y Qu, J Liu, WX Zhao, Q She, H Wu… - arxiv preprint arxiv …, 2021 - arxiv.org
In various natural language processing tasks, passage retrieval and passage re-ranking are
two key procedures in finding and ranking relevant information. Since both the two …

RocketQA: An optimized training approach to dense passage retrieval for open-domain question answering

Y Qu, Y Ding, J Liu, K Liu, R Ren, WX Zhao… - arxiv preprint arxiv …, 2020 - arxiv.org
In open-domain question answering, dense passage retrieval has become a new paradigm
to retrieve relevant passages for finding answers. Typically, the dual-encoder architecture is …

Rankt5: Fine-tuning t5 for text ranking with ranking losses

H Zhuang, Z Qin, R Jagerman, K Hui, J Ma… - Proceedings of the 46th …, 2023 - dl.acm.org
Pretrained language models such as BERT have been shown to be exceptionally effective
for text ranking. However, there are limited studies on how to leverage more powerful …

[BOK][B] Pretrained transformers for text ranking: Bert and beyond

J Lin, R Nogueira, A Yates - 2022 - books.google.com
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …

Condenser: a pre-training architecture for dense retrieval

L Gao, J Callan - arxiv preprint arxiv:2104.08253, 2021 - arxiv.org
Pre-trained Transformer language models (LM) have become go-to text representation
encoders. Prior research fine-tunes deep LMs to encode text sequences such as sentences …

COIL: Revisit exact lexical match in information retrieval with contextualized inverted list

L Gao, Z Dai, J Callan - arxiv preprint arxiv:2104.07186, 2021 - arxiv.org
Classical information retrieval systems such as BM25 rely on exact lexical match and carry
out search efficiently with inverted list index. Recent neural IR models shifts towards soft …

Improving efficient neural ranking models with cross-architecture knowledge distillation

S Hofstätter, S Althammer, M Schröder… - arxiv preprint arxiv …, 2020 - arxiv.org
Retrieval and ranking models are the backbone of many applications such as web search,
open domain QA, or text-based recommender systems. The latency of neural ranking …