Neural machine translation: A review

F Stahlberg - Journal of Artificial Intelligence Research, 2020 - jair.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …

Attention mechanism in neural networks: where it comes and where it goes

D Soydaner - Neural Computing and Applications, 2022 - Springer
A long time ago in the machine learning literature, the idea of incorporating a mechanism
inspired by the human visual system into neural networks was introduced. This idea is …

Branchformer: Parallel mlp-attention architectures to capture local and global context for speech recognition and understanding

Y Peng, S Dalmia, I Lane… - … Conference on Machine …, 2022 - proceedings.mlr.press
Conformer has proven to be effective in many speech processing tasks. It combines the
benefits of extracting local dependencies using convolutions and global dependencies …

Random feature attention

H Peng, N Pappas, D Yogatama, R Schwartz… - arxiv preprint arxiv …, 2021 - arxiv.org
Transformers are state-of-the-art models for a variety of sequence modeling tasks. At their
core is an attention function which models pairwise interactions between the inputs at every …

Transformer network for remaining useful life prediction of lithium-ion batteries

D Chen, W Hong, X Zhou - Ieee Access, 2022 - ieeexplore.ieee.org
Accurately predicting the Remaining Useful Life (RUL) of a Li-ion battery plays an important
role in managing the health and estimating the state of a battery. With the rapid development …

Attention, please! A survey of neural attention models in deep learning

A de Santana Correia, EL Colombini - Artificial Intelligence Review, 2022 - Springer
In humans, Attention is a core property of all perceptual and cognitive operations. Given our
limited ability to process competing sources, attention mechanisms select, modulate, and …

Transformers in time-series analysis: A tutorial

S Ahmed, IE Nielsen, A Tripathi, S Siddiqui… - Circuits, Systems, and …, 2023 - Springer
Transformer architectures have widespread applications, particularly in Natural Language
Processing and Computer Vision. Recently, Transformers have been employed in various …

Theoretical limitations of self-attention in neural sequence models

M Hahn - Transactions of the Association for Computational …, 2020 - direct.mit.edu
Transformers are emerging as the new workhorse of NLP, showing great success across
tasks. Unlike LSTMs, transformers process input sequences entirely through self-attention …

A survey of deep learning techniques for neural machine translation

S Yang, Y Wang, X Chu - arxiv preprint arxiv:2002.07526, 2020 - arxiv.org
In recent years, natural language processing (NLP) has got great development with deep
learning techniques. In the sub-field of machine translation, a new approach named Neural …

A survey of information extraction based on deep learning

Y Yang, Z Wu, Y Yang, S Lian, F Guo, Z Wang - Applied Sciences, 2022 - mdpi.com
As a core task and an important link in the fields of natural language understanding and
information retrieval, information extraction (IE) can structure and semanticize unstructured …