Large language models on graphs: A comprehensive survey

B **, G Liu, C Han, M Jiang, H Ji… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Large language models (LLMs), such as GPT4 and LLaMA, are creating significant
advancements in natural language processing, due to their strong text encoding/decoding …

A primer on contrastive pretraining in language processing: Methods, lessons learned, and perspectives

N Rethmeier, I Augenstein - ACM Computing Surveys, 2023 - dl.acm.org
Modern natural language processing (NLP) methods employ self-supervised pretraining
objectives such as masked language modeling to boost the performance of various …

Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing

P He, J Gao, W Chen - arxiv preprint arxiv:2111.09543, 2021 - arxiv.org
This paper presents a new pre-trained language model, DeBERTaV3, which improves the
original DeBERTa model by replacing mask language modeling (MLM) with replaced token …

Simcse: Simple contrastive learning of sentence embeddings

T Gao, X Yao, D Chen - arxiv preprint arxiv:2104.08821, 2021 - arxiv.org
This paper presents SimCSE, a simple contrastive learning framework that greatly advances
state-of-the-art sentence embeddings. We first describe an unsupervised approach, which …

Generating training data with language models: Towards zero-shot language understanding

Y Meng, J Huang, Y Zhang… - Advances in Neural …, 2022 - proceedings.neurips.cc
Pretrained language models (PLMs) have demonstrated remarkable performance in various
natural language processing tasks: Unidirectional PLMs (eg, GPT) are well known for their …

DiffCSE: Difference-based contrastive learning for sentence embeddings

YS Chuang, R Dangovski, H Luo, Y Zhang… - arxiv preprint arxiv …, 2022 - arxiv.org
We propose DiffCSE, an unsupervised contrastive learning framework for learning sentence
embeddings. DiffCSE learns sentence embeddings that are sensitive to the difference …

Contrastive learning for representation degeneration problem in sequential recommendation

R Qiu, Z Huang, H Yin, Z Wang - … conference on web search and data …, 2022 - dl.acm.org
Recent advancements of sequential deep learning models such as Transformer and BERT
have significantly facilitated the sequential recommendation. However, according to our …

A contrastive framework for neural text generation

Y Su, T Lan, Y Wang, D Yogatama… - Advances in Neural …, 2022 - proceedings.neurips.cc
Text generation is of great importance to many natural language processing applications.
However, maximization-based decoding methods (eg, beam search) of neural language …

Understanding contrastive learning requires incorporating inductive biases

N Saunshi, J Ash, S Goel, D Misra… - International …, 2022 - proceedings.mlr.press
Contrastive learning is a popular form of self-supervised learning that encourages
augmentations (views) of the same input to have more similar representations compared to …

Supporting clustering with contrastive learning

D Zhang, F Nan, X Wei, S Li, H Zhu, K McKeown… - arxiv preprint arxiv …, 2021 - arxiv.org
Unsupervised clustering aims at discovering the semantic categories of data according to
some distance measured in the representation space. However, different categories often …