PARADE: Passage Representation Aggregation forDocument Reranking

C Li, A Yates, S MacAvaney, B He, Y Sun - ACM Transactions on …, 2023 - dl.acm.org
Pre-trained transformer models, such as BERT and T5, have shown to be highly effective at
ad hoc passage and document ranking. Due to the inherent sequence length limits of these …

Inpars: Data augmentation for information retrieval using large language models

L Bonifacio, H Abonizio, M Fadaee… - arxiv preprint arxiv …, 2022 - arxiv.org
The information retrieval community has recently witnessed a revolution due to large
pretrained transformer models. Another key ingredient for this revolution was the MS …

mmarco: A multilingual version of the ms marco passage ranking dataset

L Bonifacio, V Jeronymo, HQ Abonizio… - arxiv preprint arxiv …, 2021 - arxiv.org
The MS MARCO ranking dataset has been widely used for training deep learning models for
IR tasks, achieving considerable effectiveness on diverse zero-shot scenarios. However, this …

Exploring listwise evidence reasoning with t5 for fact verification

K Jiang, R Pradeep, J Lin - … of the 59th Annual Meeting of the …, 2021 - aclanthology.org
This work explores a framework for fact verification that leverages pretrained sequence-to-
sequence transformer models for sentence selection and label prediction, two key sub-tasks …

Squeezing water from a stone: a bag of tricks for further improving cross-encoder effectiveness for reranking

R Pradeep, Y Liu, X Zhang, Y Li, A Yates… - European Conference on …, 2022 - Springer
While much recent work has demonstrated that hard negative mining can be used to train
better bi-encoder models, few have considered it in the context of cross-encoders, which are …

[PDF][PDF] No parameter left behind: How distillation and model size affect zero-shot retrieval

GM Rosa, L Bonifacio, V Jeronymo… - arxiv preprint arxiv …, 2022 - researchgate.net
Recent work has shown that small distilled language models are strong competitors to
models that are orders of magnitude larger and slower in a wide range of information …

In defense of cross-encoders for zero-shot retrieval

G Rosa, L Bonifacio, V Jeronymo, H Abonizio… - arxiv preprint arxiv …, 2022 - arxiv.org
Bi-encoders and cross-encoders are widely used in many state-of-the-art retrieval pipelines.
In this work we study the generalization ability of these two types of architectures on a wide …

Neural query synthesis and domain-specific ranking templates for multi-stage clinical trial matching

R Pradeep, Y Li, Y Wang, J Lin - … of the 45th International ACM SIGIR …, 2022 - dl.acm.org
In this work, we propose an effective multi-stage neural ranking system for the clinical trial
matching problem. First, we introduce NQS, a neural query synthesis method that leverages …

Document expansion baselines and learned sparse lexical representations for ms marco v1 and v2

X Ma, R Pradeep, R Nogueira, J Lin - Proceedings of the 45th …, 2022 - dl.acm.org
With doc2query, we train a neural sequence-to-sequence model that, given an input span of
text, predicts a natural language query that the text might answer. These predictions can be …

Billions of parameters are worth more than in-domain training data: A case study in the legal case entailment task

GM Rosa, L Bonifacio, V Jeronymo, H Abonizio… - arxiv preprint arxiv …, 2022 - arxiv.org
Recent work has shown that language models scaled to billions of parameters, such as GPT-
3, perform remarkably well in zero-shot and few-shot scenarios. In this work, we experiment …