Semantic models for the first-stage retrieval: A comprehensive review

J Guo, Y Cai, Y Fan, F Sun, R Zhang… - ACM Transactions on …, 2022 - dl.acm.org
Multi-stage ranking pipelines have been a practical solution in modern search systems,
where the first-stage retrieval is to return a subset of candidate documents and latter stages …

Information retrieval: recent advances and beyond

KA Hambarde, H Proenca - IEEE Access, 2023 - ieeexplore.ieee.org
This paper provides an extensive and thorough overview of the models and techniques
utilized in the first and second stages of the typical information retrieval processing chain …

[LIVRE][B] Pretrained transformers for text ranking: Bert and beyond

J Lin, R Nogueira, A Yates - 2022 - books.google.com
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …

Improving efficient neural ranking models with cross-architecture knowledge distillation

S Hofstätter, S Althammer, M Schröder… - arxiv preprint arxiv …, 2020 - arxiv.org
Retrieval and ranking models are the backbone of many applications such as web search,
open domain QA, or text-based recommender systems. The latency of neural ranking …

Learning passage impacts for inverted indexes

A Mallia, O Khattab, T Suel, N Tonellotto - Proceedings of the 44th …, 2021 - dl.acm.org
Neural information retrieval systems typically use a cascading pipeline, in which a first-stage
model retrieves a candidate set of documents and one or more subsequent stages re-rank …

An efficiency study for SPLADE models

C Lassance, S Clinchant - Proceedings of the 45th International ACM …, 2022 - dl.acm.org
Latency and efficiency issues are often overlooked when evaluating IR models based on
Pretrained Language Models (PLMs) in reason of multiple hardware and software testing …

Efficient document-at-a-time and score-at-a-time query evaluation for learned sparse representations

J Mackenzie, A Trotman, J Lin - ACM Transactions on Information …, 2023 - dl.acm.org
Researchers have had much recent success with ranking models based on so-called
learned sparse representations generated by transformers. One crucial advantage of this …

Faster learned sparse retrieval with guided traversal

A Mallia, J Mackenzie, T Suel, N Tonellotto - Proceedings of the 45th …, 2022 - dl.acm.org
Neural information retrieval architectures based on transformers such as BERT are able to
significantly improve system effectiveness over traditional sparse models such as BM25 …

Wacky weights in learned sparse representations and the revenge of score-at-a-time query evaluation

J Mackenzie, A Trotman, J Lin - arxiv preprint arxiv:2110.11540, 2021 - arxiv.org
Recent advances in retrieval models based on learned sparse representations generated by
transformers have led us to, once again, consider score-at-a-time query evaluation …

Efficient neural ranking using forward indexes and lightweight encoders

J Leonhardt, H Müller, K Rudra, M Khosla… - ACM Transactions on …, 2024 - dl.acm.org
Dual-encoder-based dense retrieval models have become the standard in IR. They employ
large Transformer-based language models, which are notoriously inefficient in terms of …