Large language models on graphs: A comprehensive survey
Large language models (LLMs), such as GPT4 and LLaMA, are creating significant
advancements in natural language processing, due to their strong text encoding/decoding …
advancements in natural language processing, due to their strong text encoding/decoding …
A primer on contrastive pretraining in language processing: Methods, lessons learned, and perspectives
N Rethmeier, I Augenstein - ACM Computing Surveys, 2023 - dl.acm.org
Modern natural language processing (NLP) methods employ self-supervised pretraining
objectives such as masked language modeling to boost the performance of various …
objectives such as masked language modeling to boost the performance of various …
Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing
This paper presents a new pre-trained language model, DeBERTaV3, which improves the
original DeBERTa model by replacing mask language modeling (MLM) with replaced token …
original DeBERTa model by replacing mask language modeling (MLM) with replaced token …
Simcse: Simple contrastive learning of sentence embeddings
This paper presents SimCSE, a simple contrastive learning framework that greatly advances
state-of-the-art sentence embeddings. We first describe an unsupervised approach, which …
state-of-the-art sentence embeddings. We first describe an unsupervised approach, which …
Generating training data with language models: Towards zero-shot language understanding
Pretrained language models (PLMs) have demonstrated remarkable performance in various
natural language processing tasks: Unidirectional PLMs (eg, GPT) are well known for their …
natural language processing tasks: Unidirectional PLMs (eg, GPT) are well known for their …
DiffCSE: Difference-based contrastive learning for sentence embeddings
We propose DiffCSE, an unsupervised contrastive learning framework for learning sentence
embeddings. DiffCSE learns sentence embeddings that are sensitive to the difference …
embeddings. DiffCSE learns sentence embeddings that are sensitive to the difference …
Contrastive learning for representation degeneration problem in sequential recommendation
Recent advancements of sequential deep learning models such as Transformer and BERT
have significantly facilitated the sequential recommendation. However, according to our …
have significantly facilitated the sequential recommendation. However, according to our …
A contrastive framework for neural text generation
Text generation is of great importance to many natural language processing applications.
However, maximization-based decoding methods (eg, beam search) of neural language …
However, maximization-based decoding methods (eg, beam search) of neural language …
Understanding contrastive learning requires incorporating inductive biases
Contrastive learning is a popular form of self-supervised learning that encourages
augmentations (views) of the same input to have more similar representations compared to …
augmentations (views) of the same input to have more similar representations compared to …
Supporting clustering with contrastive learning
Unsupervised clustering aims at discovering the semantic categories of data according to
some distance measured in the representation space. However, different categories often …
some distance measured in the representation space. However, different categories often …