A comprehensive survey on pretrained foundation models: A history from bert to chatgpt

C Zhou, Q Li, C Li, J Yu, Y Liu, G Wang… - International Journal of …, 2024 - Springer
Abstract Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …

[HTML][HTML] A comprehensive review on ensemble deep learning: Opportunities and challenges

A Mohammed, R Kora - Journal of King Saud University-Computer and …, 2023 - Elsevier
In machine learning, two approaches outperform traditional algorithms: ensemble learning
and deep learning. The former refers to methods that integrate multiple base models in the …

Holistic evaluation of language models

P Liang, R Bommasani, T Lee, D Tsipras… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models (LMs) are becoming the foundation for almost all major language
technologies, but their capabilities, limitations, and risks are not well understood. We present …

Simcse: Simple contrastive learning of sentence embeddings

T Gao, X Yao, D Chen - arxiv preprint arxiv:2104.08821, 2021 - arxiv.org
This paper presents SimCSE, a simple contrastive learning framework that greatly advances
state-of-the-art sentence embeddings. We first describe an unsupervised approach, which …

[PDF][PDF] Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks

N Reimers - arxiv preprint arxiv:1908.10084, 2019 - fq.pkwyx.com
BERT (Devlin et al., 2018) and RoBERTa (Liu et al., 2019) has set a new state-of-the-art
performance on sentence-pair regression tasks like semantic textual similarity (STS) …

Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity

Y Lu, M Bartolo, A Moore, S Riedel… - arxiv preprint arxiv …, 2021 - arxiv.org
When primed with only a handful of training samples, very large, pretrained language
models such as GPT-3 have shown competitive results when compared to fully-supervised …

Representation learning with contrastive predictive coding

A Oord, Y Li, O Vinyals - arxiv preprint arxiv:1807.03748, 2018 - arxiv.org
While supervised learning has enabled great progress in many applications, unsupervised
learning has not seen such widespread adoption, and remains an important and …

Deep learning--based text classification: a comprehensive review

S Minaee, N Kalchbrenner, E Cambria… - ACM computing …, 2021 - dl.acm.org
Deep learning--based models have surpassed classical machine learning--based
approaches in various text classification tasks, including sentiment analysis, news …

Language-agnostic BERT sentence embedding

F Feng, Y Yang, D Cer, N Arivazhagan… - arxiv preprint arxiv …, 2020 - arxiv.org
While BERT is an effective method for learning monolingual sentence embeddings for
semantic similarity and embedding based transfer learning (Reimers and Gurevych, 2019) …

Debiased contrastive learning

CY Chuang, J Robinson, YC Lin… - Advances in neural …, 2020 - proceedings.neurips.cc
A prominent technique for self-supervised representation learning has been to contrast
semantically similar and dissimilar pairs of samples. Without access to labels, dissimilar …