Large-scale multi-modal pre-trained models: A comprehensive survey
With the urgent demand for generalized deep models, many pre-trained big models are
proposed, such as bidirectional encoder representations (BERT), vision transformer (ViT) …
proposed, such as bidirectional encoder representations (BERT), vision transformer (ViT) …
Ammus: A survey of transformer-based pretrained models in natural language processing
KS Kalyan, A Rajasekharan, S Sangeetha - arxiv preprint arxiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …
almost every NLP task. The evolution of these models started with GPT and BERT. These …
Pre-trained models for natural language processing: A survey
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …
Sgpt: Gpt sentence embeddings for semantic search
N Muennighoff - arxiv preprint arxiv:2202.08904, 2022 - arxiv.org
Decoder transformers have continued increasing in scale reaching hundreds of billions of
parameters. Due to their scale the same decoder sets state-of-the-art results on various …
parameters. Due to their scale the same decoder sets state-of-the-art results on various …
Multimodal learning with graphs
Artificial intelligence for graphs has achieved remarkable success in modelling complex
systems, ranging from dynamic networks in biology to interacting particle systems in physics …
systems, ranging from dynamic networks in biology to interacting particle systems in physics …
Pre-trained language models in biomedical domain: A systematic survey
Pre-trained language models (PLMs) have been the de facto paradigm for most natural
language processing tasks. This also benefits the biomedical domain: researchers from …
language processing tasks. This also benefits the biomedical domain: researchers from …
mgpt: Few-shot learners go multilingual
Recent studies report that autoregressive language models can successfully solve many
NLP tasks via zero-and few-shot learning paradigms, which opens up new possibilities for …
NLP tasks via zero-and few-shot learning paradigms, which opens up new possibilities for …
[HTML][HTML] AMMU: a survey of transformer-based biomedical pretrained language models
Transformer-based pretrained language models (PLMs) have started a new era in modern
natural language processing (NLP). These models combine the power of transformers …
natural language processing (NLP). These models combine the power of transformers …
Scaling laws for language encoding models in fMRI
Abstract Representations from transformer-based unidirectional language models are
known to be effective at predicting brain responses to natural language. However, most …
known to be effective at predicting brain responses to natural language. However, most …
Limitations of transformers on clinical text classification
Bidirectional Encoder Representations from Transformers (BERT) and BERT-based
approaches are the current state-of-the-art in many natural language processing (NLP) …
approaches are the current state-of-the-art in many natural language processing (NLP) …