[PDF][PDF] 深度学**中知识蒸馏研究综述

邵仁荣, 刘宇昂, 张伟, 王骏 - 计算机学报, 2022 - 159.226.43.17
摘要在人工智能迅速发展的今天, 深度神经网络广泛应用于各个研究领域并取得了巨大的成功,
但也同样面临着诸多挑战. 首先, 为了解决复杂的问题和提高模型的训练效果 …

Temperature annealing knowledge distillation from averaged teacher

X Gu, Z Zhang, T Luo - 2022 IEEE 42nd International …, 2022 - ieeexplore.ieee.org
Despite the success of deep neural networks (DNNs) in almost every field, their deployment
on edge devices has been restricted due to the significant memory and computational …

Self-distillation with model averaging

X Gu, Z Zhang, R **, RSM Goh, T Luo - Information Sciences, 2025 - Elsevier
Abstract Knowledge distillation (KD) and model averaging (MA) are prominent techniques
for enhancing the efficiency and effectiveness of deep neural networks (DNNs). MA …

Progressively Relaxed Knowledge Distillation

X Gu, R **, M Li - 2024 International Joint Conference on …, 2024 - ieeexplore.ieee.org
Knowledge Distillation (KD) enhances the generalization ability of a student model by
transferring knowledge from a teacher model. However, literature suggests that the student …

[PDF][PDF] Hardware-Aware Co-Optimization of Deep Convolutional Neural Networks

NK Jha - 2020 - researchgate.net
The unprecedented success of deep neural networks (DNNs), especially convolutional
neural networks (CNNs), stems from its high representational power and capability to model …