Attention mechanisms in computer vision: A survey

MH Guo, TX Xu, JJ Liu, ZN Liu, PT Jiang, TJ Mu… - Computational visual …, 2022‏ - Springer
Humans can naturally and effectively find salient regions in complex scenes. Motivated by
this observation, attention mechanisms were introduced into computer vision with the aim of …

Deep learning-based 3D point cloud classification: A systematic survey and outlook

H Zhang, C Wang, S Tian, B Lu, L Zhang, X Ning, X Bai - Displays, 2023‏ - Elsevier
In recent years, point cloud representation has become one of the research hotspots in the
field of computer vision, and has been widely used in many fields, such as autonomous …

Visual attention network

MH Guo, CZ Lu, ZN Liu, MM Cheng, SM Hu - Computational visual media, 2023‏ - Springer
While originally designed for natural language processing tasks, the self-attention
mechanism has recently taken various computer vision areas by storm. However, the 2D …

Point transformer

H Zhao, L Jiang, J Jia, PHS Torr… - Proceedings of the …, 2021‏ - openaccess.thecvf.com
Self-attention networks have revolutionized natural language processing and are making
impressive strides in image analysis tasks such as image classification and object detection …

Pct: Point cloud transformer

MH Guo, JX Cai, ZN Liu, TJ Mu, RR Martin… - Computational visual …, 2021‏ - Springer
The irregular domain and lack of ordering make it challenging to design deep neural
networks for point cloud processing. This paper presents a novel framework named Point …

Self-positioning point-based transformer for point cloud understanding

J Park, S Lee, S Kim, Y **ong… - Proceedings of the IEEE …, 2023‏ - openaccess.thecvf.com
Transformers have shown superior performance on various computer vision tasks with their
capabilities to capture long-range dependencies. Despite the success, it is challenging to …

Co-scale conv-attentional image transformers

W Xu, Y Xu, T Chang, Z Tu - Proceedings of the IEEE/CVF …, 2021‏ - openaccess.thecvf.com
In this paper, we present Co-scale conv-attentional image Transformers (CoaT), a
Transformer-based image classifier equipped with co-scale and conv-attentional …

Rtformer: Efficient design for real-time semantic segmentation with transformer

J Wang, C Gou, Q Wu, H Feng, J Han… - Advances in neural …, 2022‏ - proceedings.neurips.cc
Recently, transformer-based networks have shown impressive results in semantic
segmentation. Yet for real-time semantic segmentation, pure CNN-based approaches still …

Pointcontrast: Unsupervised pre-training for 3d point cloud understanding

S **e, J Gu, D Guo, CR Qi, L Guibas… - Computer Vision–ECCV …, 2020‏ - Springer
Arguably one of the top success stories of deep learning is transfer learning. The finding that
pre-training a network on a rich source set (eg, ImageNet) can help boost performance once …

Se (3)-transformers: 3d roto-translation equivariant attention networks

F Fuchs, D Worrall, V Fischer… - Advances in neural …, 2020‏ - proceedings.neurips.cc
Abstract We introduce the SE (3)-Transformer, a variant of the self-attention module for 3D
point-clouds, which is equivariant under continuous 3D roto-translations. Equivariance is …