Riformer: Keep your vision backbone effective but removing token mixer

J Wang, S Zhang, Y Liu, T Wu, Y Yang… - Proceedings of the …, 2023 - openaccess.thecvf.com
This paper studies how to keep a vision backbone effective while removing token mixers in
its basic building blocks. Token mixers, as self-attention for vision transformers (ViTs), are …

Masked autoencoders are stronger knowledge distillers

S Lao, G Song, B Liu, Y Liu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Abstract Knowledge distillation (KD) has shown great success in improving student's
performance by mimicking the intermediate output of the high-capacity teacher in fine …

Lightweight and optimization acceleration methods for vision transformer: A review

M Chen, J Gao, W Yu - 2022 IEEE 25th International …, 2022 - ieeexplore.ieee.org
With the rapid development of technologies such as smart home, smart medical and
autonomous driving, lightweight networks play an important role in promoting the application …

UniKD: Universal Knowledge Distillation for Mimicking Homogeneous or Heterogeneous Object Detectors

S Lao, G Song, B Liu, Y Liu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Abstract Knowledge distillation (KD) has become a standard method to boost the
performance of lightweight object detectors. Most previous works are feature-based, where …

Optimizing Vision Transformers with Data-Free Knowledge Transfer

G Habib, D Singh, IA Malik, B Lall - arxiv preprint arxiv:2408.05952, 2024 - arxiv.org
The groundbreaking performance of transformers in Natural Language Processing (NLP)
tasks has led to their replacement of traditional Convolutional Neural Networks (CNNs) …

Riformer: Keep your vision backbone effective while removing token mixer

J Wang, S Zhang, Y Liu, T Wu, Y Yang, X Liu… - arxiv preprint arxiv …, 2023 - arxiv.org
This paper studies how to keep a vision backbone effective while removing token mixers in
its basic building blocks. Token mixers, as self-attention for vision transformers (ViTs), are …

An Attention-based Representation Distillation Baseline for Multi-Label Continual Learning

M Menabue, E Frascaroli, M Boschini… - arxiv preprint arxiv …, 2024 - arxiv.org
The field of Continual Learning (CL) has inspired numerous researchers over the years,
leading to increasingly advanced countermeasures to the issue of catastrophic forgetting …

Dqformer: Dynamic query transformer for lane detection

H Yang, S Lin, R Jiang, Y Lu… - ICASSP 2023-2023 IEEE …, 2023 - ieeexplore.ieee.org
Lane detection is one of the most important tasks in self-driving. The critical purpose of lane
detection is the prediction of lane shapes. Meanwhile, it is challenging and difficult to …