Spiking neural networks and their applications: A review
The past decade has witnessed the great success of deep neural networks in various
domains. However, deep neural networks are very resource-intensive in terms of energy …
domains. However, deep neural networks are very resource-intensive in terms of energy …
A survey on efficient convolutional neural networks and hardware acceleration
Over the past decade, deep-learning-based representations have demonstrated remarkable
performance in academia and industry. The learning capability of convolutional neural …
performance in academia and industry. The learning capability of convolutional neural …
Knowledge distillation: A survey
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …
especially for computer vision tasks. The great success of deep learning is mainly due to its …
Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks
L Wang, KJ Yoon - IEEE transactions on pattern analysis and …, 2021 - ieeexplore.ieee.org
Deep neural models, in recent years, have been successful in almost every field, even
solving the most complex problem statements. However, these models are huge in size with …
solving the most complex problem statements. However, these models are huge in size with …
Curriculum temperature for knowledge distillation
Most existing distillation methods ignore the flexible role of the temperature in the loss
function and fix it as a hyper-parameter that can be decided by an inefficient grid search. In …
function and fix it as a hyper-parameter that can be decided by an inefficient grid search. In …
Self-distillation: Towards efficient and compact neural networks
Remarkable achievements have been obtained by deep neural networks in the last several
years. However, the breakthrough in neural networks accuracy is always accompanied by …
years. However, the breakthrough in neural networks accuracy is always accompanied by …
A survey on green deep learning
In recent years, larger and deeper models are springing up and continuously pushing state-
of-the-art (SOTA) results across various fields like natural language processing (NLP) and …
of-the-art (SOTA) results across various fields like natural language processing (NLP) and …
Spot-adaptive knowledge distillation
Knowledge distillation (KD) has become a well established paradigm for compressing deep
neural networks. The typical way of conducting knowledge distillation is to train the student …
neural networks. The typical way of conducting knowledge distillation is to train the student …
TransKD: Transformer knowledge distillation for efficient semantic segmentation
Semantic segmentation benchmarks in the realm of autonomous driving are dominated by
large pre-trained transformers, yet their widespread adoption is impeded by substantial …
large pre-trained transformers, yet their widespread adoption is impeded by substantial …
Ernie 3.0 titan: Exploring larger-scale knowledge enhanced pre-training for language understanding and generation
Pre-trained language models have achieved state-of-the-art results in various Natural
Language Processing (NLP) tasks. GPT-3 has shown that scaling up pre-trained language …
Language Processing (NLP) tasks. GPT-3 has shown that scaling up pre-trained language …