Deep learning--based text classification: a comprehensive review

S Minaee, N Kalchbrenner, E Cambria… - ACM computing …, 2021 - dl.acm.org
Deep learning--based models have surpassed classical machine learning--based
approaches in various text classification tasks, including sentiment analysis, news …

Pre-trained models for natural language processing: A survey

X Qiu, T Sun, Y Xu, Y Shao, N Dai, X Huang - Science China …, 2020 - Springer
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …

M6-rec: Generative pretrained language models are open-ended recommender systems

Z Cui, J Ma, C Zhou, J Zhou, H Yang - arxiv preprint arxiv:2205.08084, 2022 - arxiv.org
Industrial recommender systems have been growing increasingly complex, may
involve\emph {diverse domains} such as e-commerce products and user-generated …

Efficiently teaching an effective dense retriever with balanced topic aware sampling

S Hofstätter, SC Lin, JH Yang, J Lin… - Proceedings of the 44th …, 2021 - dl.acm.org
A vital step towards the widespread adoption of neural retrieval models is their resource
efficiency throughout the training, indexing and query workflows. The neural IR community …

[KNJIGA][B] Pretrained transformers for text ranking: Bert and beyond

J Lin, R Nogueira, A Yates - 2022 - books.google.com
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …

A neural corpus indexer for document retrieval

Y Wang, Y Hou, H Wang, Z Miao… - Advances in …, 2022 - proceedings.neurips.cc
Current state-of-the-art document retrieval solutions mainly follow an index-retrieve
paradigm, where the index is hard to be directly optimized for the final retrieval target. In this …

RetroMAE: Pre-training retrieval-oriented language models via masked auto-encoder

S **ao, Z Liu, Y Shao, Z Cao - arxiv preprint arxiv:2205.12035, 2022 - arxiv.org
Despite pre-training's progress in many important NLP tasks, it remains to explore effective
pre-training strategies for dense retrieval. In this paper, we propose RetroMAE, a new …

Information retrieval: recent advances and beyond

KA Hambarde, H Proenca - IEEE Access, 2023 - ieeexplore.ieee.org
This paper provides an extensive and thorough overview of the models and techniques
utilized in the first and second stages of the typical information retrieval processing chain …

Pretrained transformers for text ranking: BERT and beyond

A Yates, R Nogueira, J Lin - Proceedings of the 14th ACM International …, 2021 - dl.acm.org
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …

Improving efficient neural ranking models with cross-architecture knowledge distillation

S Hofstätter, S Althammer, M Schröder… - arxiv preprint arxiv …, 2020 - arxiv.org
Retrieval and ranking models are the backbone of many applications such as web search,
open domain QA, or text-based recommender systems. The latency of neural ranking …