[HTML][HTML] Review of image classification algorithms based on convolutional neural networks

L Chen, S Li, Q Bai, J Yang, S Jiang, Y Miao - Remote Sensing, 2021 - mdpi.com
Image classification has always been a hot research direction in the world, and the
emergence of deep learning has promoted the development of this field. Convolutional …

Deep semantic segmentation of natural and medical images: a review

S Asgari Taghanaki, K Abhishek, JP Cohen… - Artificial intelligence …, 2021 - Springer
The semantic image segmentation task consists of classifying each pixel of an image into an
instance, where each instance corresponds to a class. This task is a part of the concept of …

Run, don't walk: chasing higher FLOPS for faster neural networks

J Chen, S Kao, H He, W Zhuo, S Wen… - Proceedings of the …, 2023 - openaccess.thecvf.com
To design fast neural networks, many works have been focusing on reducing the number of
floating-point operations (FLOPs). We observe that such reduction in FLOPs, however, does …

Mixed local channel attention for object detection

D Wan, R Lu, S Shen, T Xu, X Lang, Z Ren - Engineering Applications of …, 2023 - Elsevier
Attention mechanism, one of the most extensively utilized components in computer vision,
can assist neural networks in emphasizing significant elements and suppressing irrelevant …

Machine learning-based lung and colon cancer detection using deep feature extraction and ensemble learning

MA Talukder, MM Islam, MA Uddin, A Akhter… - Expert Systems with …, 2022 - Elsevier
Cancer is a fatal disease caused by a combination of genetic diseases and a variety of
biochemical abnormalities. Lung and colon cancer have emerged as two of the leading …

Coatnet: Marrying convolution and attention for all data sizes

Z Dai, H Liu, QV Le, M Tan - Advances in neural information …, 2021 - proceedings.neurips.cc
Transformers have attracted increasing interests in computer vision, but they still fall behind
state-of-the-art convolutional networks. In this work, we show that while Transformers tend to …

Mlp-mixer: An all-mlp architecture for vision

IO Tolstikhin, N Houlsby, A Kolesnikov… - Advances in neural …, 2021 - proceedings.neurips.cc
Abstract Convolutional Neural Networks (CNNs) are the go-to model for computer vision.
Recently, attention-based networks, such as the Vision Transformer, have also become …

Twins: Revisiting the design of spatial attention in vision transformers

X Chu, Z Tian, Y Wang, B Zhang… - Advances in neural …, 2021 - proceedings.neurips.cc
Very recently, a variety of vision transformer architectures for dense prediction tasks have
been proposed and they show that the design of spatial attention is critical to their success in …

Efficientnetv2: Smaller models and faster training

M Tan, Q Le - International conference on machine learning, 2021 - proceedings.mlr.press
This paper introduces EfficientNetV2, a new family of convolutional networks that have faster
training speed and better parameter efficiency than previous models. To develop these …

Scaling local self-attention for parameter efficient visual backbones

A Vaswani, P Ramachandran… - Proceedings of the …, 2021 - openaccess.thecvf.com
Self-attention has the promise of improving computer vision systems due to parameter-
independent scaling of receptive fields and content-dependent interactions, in contrast to …