[HTML][HTML] Pre-trained language models and their applications
Pre-trained language models have achieved striking success in natural language
processing (NLP), leading to a paradigm shift from supervised learning to pre-training …
processing (NLP), leading to a paradigm shift from supervised learning to pre-training …
[HTML][HTML] A survey of transformers
Transformers have achieved great success in many artificial intelligence fields, such as
natural language processing, computer vision, and audio processing. Therefore, it is natural …
natural language processing, computer vision, and audio processing. Therefore, it is natural …
Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation
Pre-trained models have achieved state-of-the-art results in various Natural Language
Processing (NLP) tasks. Recent works such as T5 and GPT-3 have shown that scaling up …
Processing (NLP) tasks. Recent works such as T5 and GPT-3 have shown that scaling up …
Recurrent memory transformer
Transformer-based models show their effectiveness across multiple domains and tasks. The
self-attention allows to combine information from all sequence elements into context-aware …
self-attention allows to combine information from all sequence elements into context-aware …
A survey on text classification algorithms: From text to predictions
In recent years, the exponential growth of digital documents has been met by rapid progress
in text classification techniques. Newly proposed machine learning algorithms leverage the …
in text classification techniques. Newly proposed machine learning algorithms leverage the …
Scaling transformer to 1m tokens and beyond with rmt
A major limitation for the broader scope of problems solvable by transformers is the
quadratic scaling of computational complexity with input size. In this study, we investigate …
quadratic scaling of computational complexity with input size. In this study, we investigate …
Museformer: Transformer with fine-and coarse-grained attention for music generation
Symbolic music generation aims to generate music scores automatically. A recent trend is to
use Transformer or its variants in music generation, which is, however, suboptimal, because …
use Transformer or its variants in music generation, which is, however, suboptimal, because …
Advancing transformer architecture in long-context large language models: A comprehensive survey
With the bomb ignited by ChatGPT, Transformer-based Large Language Models (LLMs)
have paved a revolutionary path toward Artificial General Intelligence (AGI) and have been …
have paved a revolutionary path toward Artificial General Intelligence (AGI) and have been …
A survey on long text modeling with transformers
Modeling long texts has been an essential technique in the field of natural language
processing (NLP). With the ever-growing number of long documents, it is important to …
processing (NLP). With the ever-growing number of long documents, it is important to …
Realistic morphology-preserving generative modelling of the brain
Medical imaging research is often limited by data scarcity and availability. Governance,
privacy concerns and the cost of acquisition all restrict access to medical imaging data …
privacy concerns and the cost of acquisition all restrict access to medical imaging data …