Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks
L Wang, KJ Yoon - IEEE transactions on pattern analysis and …, 2021 - ieeexplore.ieee.org
Deep neural models, in recent years, have been successful in almost every field, even
solving the most complex problem statements. However, these models are huge in size with …
solving the most complex problem statements. However, these models are huge in size with …
Structured pruning for deep convolutional neural networks: A survey
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …
attributed to their deeper and wider architectures, which can come with significant …
Factorizing knowledge in neural networks
In this paper, we explore a novel and ambitious knowledge-transfer task, termed Knowledge
Factorization (KF). The core idea of KF lies in the modularization and assemblability of …
Factorization (KF). The core idea of KF lies in the modularization and assemblability of …
Knowledge distillation: A survey
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …
especially for computer vision tasks. The great success of deep learning is mainly due to its …
Filtering, distillation, and hard negatives for vision-language pre-training
Vision-language models trained with contrastive learning on large-scale noisy data are
becoming increasingly popular for zero-shot recognition problems. In this paper we improve …
becoming increasingly popular for zero-shot recognition problems. In this paper we improve …
Revisiting random channel pruning for neural network compression
Channel (or 3D filter) pruning serves as an effective way to accelerate the inference of
neural networks. There has been a flurry of algorithms that try to solve this practical problem …
neural networks. There has been a flurry of algorithms that try to solve this practical problem …
Gan compression: Efficient architectures for interactive conditional gans
Abstract Conditional Generative Adversarial Networks (cGANs) have enabled controllable
image synthesis for many computer vision and graphics applications. However, recent …
image synthesis for many computer vision and graphics applications. However, recent …
Binocular mutual learning for improving few-shot classification
Z Zhou, X Qiu, J **e, J Wu… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Most of the few-shot learning methods learn to transfer knowledge from datasets with
abundant labeled data (ie, the base set). From the perspective of class space on base set …
abundant labeled data (ie, the base set). From the perspective of class space on base set …
Zero-shot knowledge transfer via adversarial belief matching
Performing knowledge transfer from a large teacher network to a smaller student is a
popular task in modern deep learning applications. However, due to growing dataset sizes …
popular task in modern deep learning applications. However, due to growing dataset sizes …
Thieves on sesame street! model extraction of bert-based apis
We study the problem of model extraction in natural language processing, in which an
adversary with only query access to a victim model attempts to reconstruct a local copy of …
adversary with only query access to a victim model attempts to reconstruct a local copy of …