A review on the attention mechanism of deep learning

Z Niu, G Zhong, H Yu - Neurocomputing, 2021 - Elsevier
Attention has arguably become one of the most important concepts in the deep learning
field. It is inspired by the biological systems of humans that tend to focus on the distinctive …

A survey of the usages of deep learning for natural language processing

DW Otter, JR Medina, JK Kalita - IEEE transactions on neural …, 2020 - ieeexplore.ieee.org
Over the last several years, the field of natural language processing has been propelled
forward by an explosion in the use of deep learning models. This article provides a brief …

Lost in the middle: How language models use long contexts

NF Liu, K Lin, J Hewitt, A Paranjape… - Transactions of the …, 2024 - direct.mit.edu
While recent language models have the ability to take long contexts as input, relatively little
is known about how well they use longer context. We analyze the performance of language …

Artificial intelligence for the metaverse: A survey

T Huynh-The, QV Pham, XQ Pham, TT Nguyen… - … Applications of Artificial …, 2023 - Elsevier
Along with the massive growth of the Internet from the 1990s until now, various innovative
technologies have been created to bring users breathtaking experiences with more virtual …

A general survey on attention mechanisms in deep learning

G Brauwers, F Frasincar - IEEE Transactions on Knowledge …, 2021 - ieeexplore.ieee.org
Attention is an important mechanism that can be employed for a variety of deep learning
models across many different domains and tasks. This survey provides an overview of the …

Long range arena: A benchmark for efficient transformers

Y Tay, M Dehghani, S Abnar, Y Shen, D Bahri… - arxiv preprint arxiv …, 2020 - arxiv.org
Transformers do not scale very well to long sequence lengths largely because of quadratic
self-attention complexity. In the recent months, a wide spectrum of efficient, fast Transformers …

Hopfield networks is all you need

H Ramsauer, B Schäfl, J Lehner, P Seidl… - arxiv preprint arxiv …, 2020 - arxiv.org
We introduce a modern Hopfield network with continuous states and a corresponding
update rule. The new Hopfield network can store exponentially (with the dimension of the …

EEG-based emotion recognition via channel-wise attention and self attention

W Tao, C Li, R Song, J Cheng, Y Liu… - IEEE Transactions …, 2020 - ieeexplore.ieee.org
Emotion recognition based on electroencephalography (EEG) is a significant task in the
brain-computer interface field. Recently, many deep learning-based emotion recognition …

Attention, please! A survey of neural attention models in deep learning

A de Santana Correia, EL Colombini - Artificial Intelligence Review, 2022 - Springer
In humans, Attention is a core property of all perceptual and cognitive operations. Given our
limited ability to process competing sources, attention mechanisms select, modulate, and …

Bidirectional LSTM with attention mechanism and convolutional layer for text classification

G Liu, J Guo - Neurocomputing, 2019 - Elsevier
Neural network models have been widely used in the field of natural language processing
(NLP). Recurrent neural networks (RNNs), which have the ability to process sequences of …