Lightweight deep learning for resource-constrained environments: A survey

HI Liu, M Galindo, H **e, LK Wong, HH Shuai… - ACM Computing …, 2024 - dl.acm.org
Over the past decade, the dominance of deep learning has prevailed across various
domains of artificial intelligence, including natural language processing, computer vision …

A survey of design and optimization for systolic array-based dnn accelerators

R Xu, S Ma, Y Guo, D Li - ACM Computing Surveys, 2023 - dl.acm.org
In recent years, it has been witnessed that the systolic array is a successful architecture for
DNN hardware accelerators. However, the design of systolic arrays also encountered many …

Efficientvit: Memory efficient vision transformer with cascaded group attention

X Liu, H Peng, N Zheng, Y Yang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Vision transformers have shown great success due to their high model capabilities.
However, their remarkable performance is accompanied by heavy computation costs, which …

Inceptionnext: When inception meets convnext

W Yu, P Zhou, S Yan, X Wang - Proceedings of the IEEE/cvf …, 2024 - openaccess.thecvf.com
Inspired by the long-range modeling ability of ViTs large-kernel convolutions are widely
studied and adopted recently to enlarge the receptive field and improve model performance …

Scale-aware modulation meet transformer

W Lin, Z Wu, J Chen, J Huang… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
This paper presents a new vision Transformer, Scale Aware Modulation Transformer (SMT),
that can handle various downstream tasks efficiently by combining the convolutional network …

Mobileone: An improved one millisecond mobile backbone

PKA Vasu, J Gabriel, J Zhu, O Tuzel… - Proceedings of the …, 2023 - openaccess.thecvf.com
Efficient neural network backbones for mobile devices are often optimized for metrics such
as FLOPs or parameter count. However, these metrics may not correlate well with latency of …

Mobile-former: Bridging mobilenet and transformer

Y Chen, X Dai, D Chen, M Liu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Abstract We present Mobile-Former, a parallel design of MobileNet and transformer with a
two-way bridge in between. This structure leverages the advantages of MobileNet at local …

Conv2former: A simple transformer-style convnet for visual recognition

Q Hou, CZ Lu, MM Cheng… - IEEE transactions on …, 2024 - ieeexplore.ieee.org
Vision Transformers have been the most popular network architecture in visual recognition
recently due to the strong ability of encode global information. However, its high …

Edgevits: Competing light-weight cnns on mobile devices with vision transformers

J Pan, A Bulat, F Tan, X Zhu, L Dudziak, H Li… - European conference on …, 2022 - Springer
Self-attention based models such as vision transformers (ViTs) have emerged as a very
competitive architecture alternative to convolutional neural networks (CNNs) in computer …

Efficientnetv2: Smaller models and faster training

M Tan, Q Le - International conference on machine learning, 2021 - proceedings.mlr.press
This paper introduces EfficientNetV2, a new family of convolutional networks that have faster
training speed and better parameter efficiency than previous models. To develop these …