[HTML][HTML] A survey of transformers

T Lin, Y Wang, X Liu, X Qiu - AI open, 2022 - Elsevier
Transformers have achieved great success in many artificial intelligence fields, such as
natural language processing, computer vision, and audio processing. Therefore, it is natural …

A Complete Process of Text Classification System Using State‐of‐the‐Art NLP Models

V Dogra, S Verma, Kavita, P Chatterjee… - Computational …, 2022 - Wiley Online Library
With the rapid advancement of information technology, online information has been
exponentially growing day by day, especially in the form of text documents such as news …

Big bird: Transformers for longer sequences

M Zaheer, G Guruganesh, KA Dubey… - Advances in neural …, 2020 - proceedings.neurips.cc
Transformers-based models, such as BERT, have been one of the most successful deep
learning models for NLP. Unfortunately, one of their core limitations is the quadratic …

Longformer: The long-document transformer

I Beltagy, ME Peters, A Cohan - arxiv preprint arxiv:2004.05150, 2020 - arxiv.org
Transformer-based models are unable to process long sequences due to their self-attention
operation, which scales quadratically with the sequence length. To address this limitation …

Perceiver: General perception with iterative attention

A Jaegle, F Gimeno, A Brock… - International …, 2021 - proceedings.mlr.press
Biological systems understand the world by simultaneously processing high-dimensional
inputs from modalities as diverse as vision, audition, touch, proprioception, etc. The …

ETC: Encoding long and structured inputs in transformers

J Ainslie, S Ontanon, C Alberti, V Cvicek… - arxiv preprint arxiv …, 2020 - arxiv.org
Transformer models have advanced the state of the art in many Natural Language
Processing (NLP) tasks. In this paper, we present a new Transformer architecture, Extended …

Artificial intelligence in the battle against coronavirus (COVID-19): a survey and future research directions

TT Nguyen, QVH Nguyen, DT Nguyen, S Yang… - arxiv preprint arxiv …, 2020 - arxiv.org
Artificial intelligence (AI) has been applied widely in our daily lives in a variety of ways with
numerous success stories. AI has also contributed to dealing with the coronavirus disease …

Long-short transformer: Efficient transformers for language and vision

C Zhu, W **, C **ao, M Shoeybi… - Advances in neural …, 2021 - proceedings.neurips.cc
Transformers have achieved success in both language and vision domains. However, it is
prohibitively expensive to scale them to long sequences such as long documents or high …

Museformer: Transformer with fine-and coarse-grained attention for music generation

B Yu, P Lu, R Wang, W Hu, X Tan… - Advances in …, 2022 - proceedings.neurips.cc
Symbolic music generation aims to generate music scores automatically. A recent trend is to
use Transformer or its variants in music generation, which is, however, suboptimal, because …

Long-range transformers for dynamic spatiotemporal forecasting

J Grigsby, Z Wang, N Nguyen, Y Qi - arxiv preprint arxiv:2109.12218, 2021 - arxiv.org
Multivariate time series forecasting focuses on predicting future values based on historical
context. State-of-the-art sequence-to-sequence models rely on neural attention between …