A comprehensive survey on pretrained foundation models: A history from bert to chatgpt
Abstract Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …
[HTML][HTML] A comprehensive review on ensemble deep learning: Opportunities and challenges
In machine learning, two approaches outperform traditional algorithms: ensemble learning
and deep learning. The former refers to methods that integrate multiple base models in the …
and deep learning. The former refers to methods that integrate multiple base models in the …
Holistic evaluation of language models
Language models (LMs) are becoming the foundation for almost all major language
technologies, but their capabilities, limitations, and risks are not well understood. We present …
technologies, but their capabilities, limitations, and risks are not well understood. We present …
Simcse: Simple contrastive learning of sentence embeddings
This paper presents SimCSE, a simple contrastive learning framework that greatly advances
state-of-the-art sentence embeddings. We first describe an unsupervised approach, which …
state-of-the-art sentence embeddings. We first describe an unsupervised approach, which …
[PDF][PDF] Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
N Reimers - arxiv preprint arxiv:1908.10084, 2019 - fq.pkwyx.com
BERT (Devlin et al., 2018) and RoBERTa (Liu et al., 2019) has set a new state-of-the-art
performance on sentence-pair regression tasks like semantic textual similarity (STS) …
performance on sentence-pair regression tasks like semantic textual similarity (STS) …
Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity
When primed with only a handful of training samples, very large, pretrained language
models such as GPT-3 have shown competitive results when compared to fully-supervised …
models such as GPT-3 have shown competitive results when compared to fully-supervised …
Representation learning with contrastive predictive coding
While supervised learning has enabled great progress in many applications, unsupervised
learning has not seen such widespread adoption, and remains an important and …
learning has not seen such widespread adoption, and remains an important and …
Deep learning--based text classification: a comprehensive review
Deep learning--based models have surpassed classical machine learning--based
approaches in various text classification tasks, including sentiment analysis, news …
approaches in various text classification tasks, including sentiment analysis, news …
Language-agnostic BERT sentence embedding
While BERT is an effective method for learning monolingual sentence embeddings for
semantic similarity and embedding based transfer learning (Reimers and Gurevych, 2019) …
semantic similarity and embedding based transfer learning (Reimers and Gurevych, 2019) …
Debiased contrastive learning
A prominent technique for self-supervised representation learning has been to contrast
semantically similar and dissimilar pairs of samples. Without access to labels, dissimilar …
semantically similar and dissimilar pairs of samples. Without access to labels, dissimilar …