RetroMAE: Pre-training retrieval-oriented language models via masked auto-encoder
Despite pre-training's progress in many important NLP tasks, it remains to explore effective
pre-training strategies for dense retrieval. In this paper, we propose RetroMAE, a new …
pre-training strategies for dense retrieval. In this paper, we propose RetroMAE, a new …
Exploring the benefits of training expert language models over instruction tuning
Abstract Recently, Language Models (LMs) instruction-tuned on multiple tasks, also known
as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen …
as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen …
Angle-optimized text embeddings
High-quality text embedding is pivotal in improving semantic textual similarity (STS) tasks,
which are crucial components in Large Language Model (LLM) applications. However, a …
which are crucial components in Large Language Model (LLM) applications. However, a …
Simlm: Pre-training with representation bottleneck for dense passage retrieval
In this paper, we propose SimLM (Similarity matching with Language Model pre-training), a
simple yet effective pre-training method for dense passage retrieval. It employs a simple …
simple yet effective pre-training method for dense passage retrieval. It employs a simple …
Improving contrastive learning of sentence embeddings from ai feedback
Contrastive learning has become a popular approach in natural language processing,
particularly for the learning of sentence embeddings. However, the discrete nature of natural …
particularly for the learning of sentence embeddings. However, the discrete nature of natural …
Rasa: Relation and sensitivity aware representation learning for text-based person search
Text-based person search aims to retrieve the specified person images given a textual
description. The key to tackling such a challenging task is to learn powerful multi-modal …
description. The key to tackling such a challenging task is to learn powerful multi-modal …
Scaling sentence embeddings with large language models
Large language models (LLMs) have recently garnered significant interest. With in-context
learning, LLMs achieve impressive results in various natural language tasks. However, the …
learning, LLMs achieve impressive results in various natural language tasks. However, the …
Infocse: Information-aggregated contrastive learning of sentence embeddings
Contrastive learning has been extensively studied in sentence embedding learning, which
assumes that the embeddings of different views of the same sentence are closer. The …
assumes that the embeddings of different views of the same sentence are closer. The …
CLSEP: Contrastive learning of sentence embedding with prompt
Sentence embedding, which aims to learn an effective representation of the sentence, is
beneficial for downstream tasks. By utilizing contrastive learning, most recent sentence …
beneficial for downstream tasks. By utilizing contrastive learning, most recent sentence …
Equivariant contrastive learning for sequential recommendation
Contrastive learning (CL) benefits the training of sequential recommendation models with
informative self-supervision signals. Existing solutions apply general sequential data …
informative self-supervision signals. Existing solutions apply general sequential data …