A comprehensive survey on applications of transformers for deep learning tasks

S Islam, H Elmekki, A Elsebai, J Bentahar… - Expert Systems with …, 2024 - Elsevier
Abstract Transformers are Deep Neural Networks (DNN) that utilize a self-attention
mechanism to capture contextual relationships within sequential data. Unlike traditional …

Attention mechanism in neural networks: where it comes and where it goes

D Soydaner - Neural Computing and Applications, 2022 - Springer
A long time ago in the machine learning literature, the idea of incorporating a mechanism
inspired by the human visual system into neural networks was introduced. This idea is …

Multiattention network for semantic segmentation of fine-resolution remote sensing images

R Li, S Zheng, C Zhang, C Duan, J Su… - … on Geoscience and …, 2021 - ieeexplore.ieee.org
Semantic segmentation of remote sensing images plays an important role in a wide range of
applications, including land resource management, biosphere monitoring, and urban …

Matnet: Motion-attentive transition network for zero-shot video object segmentation

T Zhou, J Li, S Wang, R Tao… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
In this paper, we present a novel end-to-end learning neural network, ie, MATNet, for zero-
shot video object segmentation (ZVOS). Motivated by the human visual attention behavior …

Deep visual attention prediction

W Wang, J Shen - IEEE Transactions on Image Processing, 2017 - ieeexplore.ieee.org
In this paper, we aim to predict human eye fixation with view-free scenes based on an end-to-
end deep learning architecture. Although convolutional neural networks (CNNs) have made …

Revisiting video saliency prediction in the deep learning era

W Wang, J Shen, J **e, MM Cheng… - IEEE transactions on …, 2019 - ieeexplore.ieee.org
Predicting where people look in static scenes, aka visual saliency, has received significant
research interest recently. However, relatively less effort has been spent in understanding …

What do different evaluation metrics tell us about saliency models?

Z Bylinskii, T Judd, A Oliva, A Torralba… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
How best to evaluate a saliency model's ability to predict where humans look in images is an
open research question. The choice of evaluation metric depends on how saliency is …

Learning unsupervised video object segmentation through visual attention

W Wang, H Song, S Zhao, J Shen… - Proceedings of the …, 2019 - openaccess.thecvf.com
This paper conducts a systematic study on the role of visual attention in Unsupervised Video
Object Segmentation (UVOS) tasks. By elaborately annotating three popular video …

Revisiting video saliency: A large-scale benchmark and a new model

W Wang, J Shen, F Guo… - Proceedings of the …, 2018 - openaccess.thecvf.com
In this work, we contribute to video saliency research in two ways. First, we introduce a new
benchmark for predicting human eye movements during dynamic scene free-viewing, which …

State-of-the-art in visual attention modeling

A Borji, L Itti - IEEE transactions on pattern analysis and …, 2012 - ieeexplore.ieee.org
Modeling visual attention-particularly stimulus-driven, saliency-based attention-has been a
very active research area over the past 25 years. Many different models of attention are now …