Reciprocal teacher-student learning via forward and feedback knowledge distillation

J Gou, Y Chen, B Yu, J Liu, L Du… - IEEE transactions on …, 2024 - ieeexplore.ieee.org
Knowledge distillation (KD) is a prevalent model compression technique in deep learning,
aiming to leverage knowledge from a large teacher model to enhance the training of a …

Logit standardization in knowledge distillation

S Sun, W Ren, J Li, R Wang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Knowledge distillation involves transferring soft labels from a teacher to a student
using a shared temperature-based softmax function. However the assumption of a shared …

Promptkd: Unsupervised prompt distillation for vision-language models

Z Li, X Li, X Fu, X Zhang, W Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Prompt learning has emerged as a valuable technique in enhancing vision-language
models (VLMs) such as CLIP for downstream tasks in specific domains. Existing work mainly …

CrossKD: Cross-head knowledge distillation for object detection

J Wang, Y Chen, Z Zheng, X Li… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Knowledge Distillation (KD) has been validated as an effective model compression
technique for learning compact object detectors. Existing state-of-the-art KD methods for …

Dual teachers for self-knowledge distillation

Z Li, X Li, L Yang, R Song, J Yang, Z Pan - Pattern Recognition, 2024 - Elsevier
We introduce an efficient self-knowledge distillation framework, Dual Teachers for Self-
Knowledge Distillation (DTSKD), where the student receives self-supervisions by dual …

Clip-kd: An empirical study of clip model distillation

C Yang, Z An, L Huang, J Bi, X Yu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Contrastive Language-Image Pre-training (CLIP) has become a promising
language-supervised visual pre-training framework. This paper aims to distill small CLIP …

Amd: Automatic multi-step distillation of large-scale vision models

C Han, Q Wang, SA Dianat, M Rabbani… - … on Computer Vision, 2024 - Springer
Transformer-based architectures have become the de-facto standard models for diverse
vision tasks owing to their superior performance. As the size of these transformer-based …

Towards Federated Large Language Models: Motivations, Methods, and Future Directions

Y Cheng, W Zhang, Z Zhang, C Zhang… - … Surveys & Tutorials, 2024 - ieeexplore.ieee.org
Large Language Models (LLMs), such as LLaMA and GPT-4, have transformed the
paradigm of natural language comprehension and generation. Despite their impressive …

Cascade prompt learning for vision-language model adaptation

G Wu, X Zhang, Z Li, Z Chen, J Liang, J Yang… - European Conference on …, 2024 - Springer
Prompt learning has surfaced as an effective approach to enhance the performance of
Vision-Language Models (VLMs) like CLIP when applied to downstream tasks. However …

Cross-domain visual prompting with spatial proximity knowledge distillation for histological image classification

X Li, G Huang, L Cheng, G Zhong, W Liu… - Journal of Biomedical …, 2024 - Elsevier
Objective: Histological classification is a challenging task due to the diverse appearances,
unpredictable variations, and blurry edges of histological tissues. Recently, many …