Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks

L Wang, KJ Yoon - IEEE transactions on pattern analysis and …, 2021 - ieeexplore.ieee.org
Deep neural models, in recent years, have been successful in almost every field, even
solving the most complex problem statements. However, these models are huge in size with …

Structured pruning for deep convolutional neural networks: A survey

Y He, L **ao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

Factorizing knowledge in neural networks

X Yang, J Ye, X Wang - European Conference on Computer Vision, 2022 - Springer
In this paper, we explore a novel and ambitious knowledge-transfer task, termed Knowledge
Factorization (KF). The core idea of KF lies in the modularization and assemblability of …

Knowledge distillation: A survey

J Gou, B Yu, SJ Maybank, D Tao - International Journal of Computer Vision, 2021 - Springer
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …

Filtering, distillation, and hard negatives for vision-language pre-training

F Radenovic, A Dubey, A Kadian… - Proceedings of the …, 2023 - openaccess.thecvf.com
Vision-language models trained with contrastive learning on large-scale noisy data are
becoming increasingly popular for zero-shot recognition problems. In this paper we improve …

Revisiting random channel pruning for neural network compression

Y Li, K Adamczewski, W Li, S Gu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Channel (or 3D filter) pruning serves as an effective way to accelerate the inference of
neural networks. There has been a flurry of algorithms that try to solve this practical problem …

Gan compression: Efficient architectures for interactive conditional gans

M Li, J Lin, Y Ding, Z Liu, JY Zhu… - Proceedings of the …, 2020 - openaccess.thecvf.com
Abstract Conditional Generative Adversarial Networks (cGANs) have enabled controllable
image synthesis for many computer vision and graphics applications. However, recent …

Binocular mutual learning for improving few-shot classification

Z Zhou, X Qiu, J **e, J Wu… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Most of the few-shot learning methods learn to transfer knowledge from datasets with
abundant labeled data (ie, the base set). From the perspective of class space on base set …

Zero-shot knowledge transfer via adversarial belief matching

P Micaelli, AJ Storkey - Advances in Neural Information …, 2019 - proceedings.neurips.cc
Performing knowledge transfer from a large teacher network to a smaller student is a
popular task in modern deep learning applications. However, due to growing dataset sizes …

Thieves on sesame street! model extraction of bert-based apis

K Krishna, GS Tomar, AP Parikh, N Papernot… - arxiv preprint arxiv …, 2019 - arxiv.org
We study the problem of model extraction in natural language processing, in which an
adversary with only query access to a victim model attempts to reconstruct a local copy of …