Scaling local self-attention for parameter efficient visual backbones

A Vaswani, P Ramachandran… - Proceedings of the …, 2021 - openaccess.thecvf.com
Self-attention has the promise of improving computer vision systems due to parameter-
independent scaling of receptive fields and content-dependent interactions, in contrast to …

Autoformer: Searching transformers for visual recognition

M Chen, H Peng, J Fu, H Ling - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Recently, pure transformer-based models have shown great potentials for vision tasks such
as image classification and detection. However, the design of transformer networks is …

Flexivit: One model for all patch sizes

L Beyer, P Izmailov, A Kolesnikov… - Proceedings of the …, 2023 - openaccess.thecvf.com
Vision Transformers convert images to sequences by slicing them into patches. The size of
these patches controls a speed/accuracy tradeoff, with smaller patches leading to higher …

AutoML: A survey of the state-of-the-art

X He, K Zhao, X Chu - Knowledge-based systems, 2021 - Elsevier
Deep learning (DL) techniques have obtained remarkable achievements on various tasks,
such as image recognition, object detection, and language modeling. However, building a …

Neural architecture search on imagenet in four gpu hours: A theoretically inspired perspective

W Chen, X Gong, Z Wang - arxiv preprint arxiv:2102.11535, 2021 - arxiv.org
Neural Architecture Search (NAS) has been explosively studied to automate the discovery of
top-performer neural networks. Current works require heavy training of supernet or intensive …

Fairnas: Rethinking evaluation fairness of weight sharing neural architecture search

X Chu, B Zhang, R Xu - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
One of the most critical problems in weight-sharing neural architecture search is the
evaluation of candidate models within a predefined search space. In practice, a one-shot …

Dynamic slimmable network

C Li, G Wang, B Wang, X Liang… - Proceedings of the …, 2021 - openaccess.thecvf.com
Current dynamic networks and dynamic pruning methods have shown their promising
capability in reducing theoretical computation complexity. However, dynamic sparse …