Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks

L Wang, KJ Yoon - IEEE transactions on pattern analysis and …, 2021 - ieeexplore.ieee.org
Deep neural models, in recent years, have been successful in almost every field, even
solving the most complex problem statements. However, these models are huge in size with …

Learn from model beyond fine-tuning: A survey

H Zheng, L Shen, A Tang, Y Luo, H Hu, B Du… - arxiv preprint arxiv …, 2023 - arxiv.org
Foundation models (FM) have demonstrated remarkable performance across a wide range
of tasks (especially in the fields of natural language processing and computer vision) …

Data-free knowledge distillation for heterogeneous federated learning

Z Zhu, J Hong, J Zhou - International conference on machine …, 2021 - proceedings.mlr.press
Federated Learning (FL) is a decentralized machine-learning paradigm, in which a global
server iteratively averages the model parameters of local users without accessing their data …

Knowledge distillation: A survey

J Gou, B Yu, SJ Maybank, D Tao - International Journal of Computer Vision, 2021 - Springer
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …

Source-free domain adaptation for semantic segmentation

Y Liu, W Zhang, J Wang - … of the IEEE/CVF conference on …, 2021 - openaccess.thecvf.com
Abstract Unsupervised Domain Adaptation (UDA) can tackle the challenge that
convolutional neural network (CNN)-based approaches for semantic segmentation heavily …

Dafkd: Domain-aware federated knowledge distillation

H Wang, Y Li, W Xu, R Li, Y Zhan… - Proceedings of the …, 2023 - openaccess.thecvf.com
Federated Distillation (FD) has recently attracted increasing attention for its efficiency in
aggregating multiple diverse local models trained from statistically heterogeneous data of …

Data-free knowledge distillation via feature exchange and activation region constraint

S Yu, J Chen, H Han, S Jiang - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Despite the tremendous progress on data-free knowledge distillation (DFKD) based on
synthetic data generation, there are still limitations in diverse and efficient data synthesis. It …

Generative low-bitwidth data free quantization

S Xu, H Li, B Zhuang, J Liu, J Cao, C Liang… - Computer Vision–ECCV …, 2020 - Springer
Neural network quantization is an effective way to compress deep models and improve their
execution latency and energy efficiency, so that they can be deployed on mobile or …

Data-free network quantization with adversarial knowledge distillation

Y Choi, J Choi, M El-Khamy… - Proceedings of the IEEE …, 2020 - openaccess.thecvf.com
Network quantization is an essential procedure in deep learning for development of efficient
fixed-point inference models on mobile or edge platforms. However, as datasets grow larger …

Self-distillation as instance-specific label smoothing

Z Zhang, M Sabuncu - Advances in Neural Information …, 2020 - proceedings.neurips.cc
It has been recently demonstrated that multi-generational self-distillation can improve
generalization. Despite this intriguing observation, reasons for the enhancement remain …