Visual tuning
Fine-tuning visual models has been widely shown promising performance on many
downstream visual tasks. With the surprising development of pre-trained visual foundation …
downstream visual tasks. With the surprising development of pre-trained visual foundation …
Knowledge distillation from a stronger teacher
Unlike existing knowledge distillation methods focus on the baseline settings, where the
teacher models and training strategies are not that strong and competing as state-of-the-art …
teacher models and training strategies are not that strong and competing as state-of-the-art …
Focal and global knowledge distillation for detectors
Abstract Knowledge distillation has been applied to image classification successfully.
However, object detection is much more sophisticated and most knowledge distillation …
However, object detection is much more sophisticated and most knowledge distillation …
Masked generative distillation
Abstract Knowledge distillation has been applied to various tasks successfully. The current
distillation algorithm usually improves students' performance by imitating the output of the …
distillation algorithm usually improves students' performance by imitating the output of the …
Knowledge diffusion for distillation
The representation gap between teacher and student is an emerging topic in knowledge
distillation (KD). To reduce the gap and improve the performance, current methods often …
distillation (KD). To reduce the gap and improve the performance, current methods often …
Knowledge distillation via the target-aware transformer
Abstract Knowledge distillation becomes a de facto standard to improve the performance of
small neural networks. Most of the previous works propose to regress the representational …
small neural networks. Most of the previous works propose to regress the representational …
Channel-wise knowledge distillation for dense prediction
Abstract Knowledge distillation (KD) has been proven a simple and effective tool for training
compact dense prediction models. Lightweight student networks are trained by extra …
compact dense prediction models. Lightweight student networks are trained by extra …
When object detection meets knowledge distillation: A survey
Object detection (OD) is a crucial computer vision task that has seen the development of
many algorithms and models over the years. While the performance of current OD models …
many algorithms and models over the years. While the performance of current OD models …
Automated knowledge distillation via monte carlo tree search
In this paper, we present Auto-KD, the first automated search framework for optimal
knowledge distillation design. Traditional distillation techniques typically require handcrafted …
knowledge distillation design. Traditional distillation techniques typically require handcrafted …
Consistency-and dependence-guided knowledge distillation for object detection in remote sensing images
As one of the challenging tasks in the remote sensing (RS), object detection has been
successfully applied in many fields. Convolution neural network (CNN) has recently …
successfully applied in many fields. Convolution neural network (CNN) has recently …