Neural machine translation: A review
F Stahlberg - Journal of Artificial Intelligence Research, 2020 - jair.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …
natural language into another, has experienced a major paradigm shift in recent years …
Attention mechanism in neural networks: where it comes and where it goes
D Soydaner - Neural Computing and Applications, 2022 - Springer
A long time ago in the machine learning literature, the idea of incorporating a mechanism
inspired by the human visual system into neural networks was introduced. This idea is …
inspired by the human visual system into neural networks was introduced. This idea is …
Branchformer: Parallel mlp-attention architectures to capture local and global context for speech recognition and understanding
Conformer has proven to be effective in many speech processing tasks. It combines the
benefits of extracting local dependencies using convolutions and global dependencies …
benefits of extracting local dependencies using convolutions and global dependencies …
Random feature attention
Transformers are state-of-the-art models for a variety of sequence modeling tasks. At their
core is an attention function which models pairwise interactions between the inputs at every …
core is an attention function which models pairwise interactions between the inputs at every …
Transformer network for remaining useful life prediction of lithium-ion batteries
D Chen, W Hong, X Zhou - Ieee Access, 2022 - ieeexplore.ieee.org
Accurately predicting the Remaining Useful Life (RUL) of a Li-ion battery plays an important
role in managing the health and estimating the state of a battery. With the rapid development …
role in managing the health and estimating the state of a battery. With the rapid development …
Attention, please! A survey of neural attention models in deep learning
In humans, Attention is a core property of all perceptual and cognitive operations. Given our
limited ability to process competing sources, attention mechanisms select, modulate, and …
limited ability to process competing sources, attention mechanisms select, modulate, and …
Transformers in time-series analysis: A tutorial
Transformer architectures have widespread applications, particularly in Natural Language
Processing and Computer Vision. Recently, Transformers have been employed in various …
Processing and Computer Vision. Recently, Transformers have been employed in various …
Theoretical limitations of self-attention in neural sequence models
M Hahn - Transactions of the Association for Computational …, 2020 - direct.mit.edu
Transformers are emerging as the new workhorse of NLP, showing great success across
tasks. Unlike LSTMs, transformers process input sequences entirely through self-attention …
tasks. Unlike LSTMs, transformers process input sequences entirely through self-attention …
A survey of deep learning techniques for neural machine translation
In recent years, natural language processing (NLP) has got great development with deep
learning techniques. In the sub-field of machine translation, a new approach named Neural …
learning techniques. In the sub-field of machine translation, a new approach named Neural …
A survey of information extraction based on deep learning
Y Yang, Z Wu, Y Yang, S Lian, F Guo, Z Wang - Applied Sciences, 2022 - mdpi.com
As a core task and an important link in the fields of natural language understanding and
information retrieval, information extraction (IE) can structure and semanticize unstructured …
information retrieval, information extraction (IE) can structure and semanticize unstructured …