Large-scale multi-modal pre-trained models: A comprehensive survey

X Wang, G Chen, G Qian, P Gao, XY Wei… - Machine Intelligence …, 2023 - Springer
With the urgent demand for generalized deep models, many pre-trained big models are
proposed, such as bidirectional encoder representations (BERT), vision transformer (ViT) …

Ammus: A survey of transformer-based pretrained models in natural language processing

KS Kalyan, A Rajasekharan, S Sangeetha - arxiv preprint arxiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …

Pre-trained models for natural language processing: A survey

X Qiu, T Sun, Y Xu, Y Shao, N Dai, X Huang - Science China …, 2020 - Springer
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …

Sgpt: Gpt sentence embeddings for semantic search

N Muennighoff - arxiv preprint arxiv:2202.08904, 2022 - arxiv.org
Decoder transformers have continued increasing in scale reaching hundreds of billions of
parameters. Due to their scale the same decoder sets state-of-the-art results on various …

Multimodal learning with graphs

Y Ektefaie, G Dasoulas, A Noori, M Farhat… - Nature Machine …, 2023 - nature.com
Artificial intelligence for graphs has achieved remarkable success in modelling complex
systems, ranging from dynamic networks in biology to interacting particle systems in physics …

Pre-trained language models in biomedical domain: A systematic survey

B Wang, Q **e, J Pei, Z Chen, P Tiwari, Z Li… - ACM Computing …, 2023 - dl.acm.org
Pre-trained language models (PLMs) have been the de facto paradigm for most natural
language processing tasks. This also benefits the biomedical domain: researchers from …

mgpt: Few-shot learners go multilingual

O Shliazhko, A Fenogenova, M Tikhonova… - arxiv preprint arxiv …, 2022 - arxiv.org
Recent studies report that autoregressive language models can successfully solve many
NLP tasks via zero-and few-shot learning paradigms, which opens up new possibilities for …

[HTML][HTML] AMMU: a survey of transformer-based biomedical pretrained language models

KS Kalyan, A Rajasekharan, S Sangeetha - Journal of biomedical …, 2022 - Elsevier
Transformer-based pretrained language models (PLMs) have started a new era in modern
natural language processing (NLP). These models combine the power of transformers …

Scaling laws for language encoding models in fMRI

R Antonello, A Vaidya, A Huth - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Representations from transformer-based unidirectional language models are
known to be effective at predicting brain responses to natural language. However, most …

Limitations of transformers on clinical text classification

S Gao, M Alawad, MT Young, J Gounley… - IEEE journal of …, 2021 - ieeexplore.ieee.org
Bidirectional Encoder Representations from Transformers (BERT) and BERT-based
approaches are the current state-of-the-art in many natural language processing (NLP) …