Spiking neural networks and their applications: A review

K Yamazaki, VK Vo-Ho, D Bulsara, N Le - Brain Sciences, 2022 - mdpi.com
The past decade has witnessed the great success of deep neural networks in various
domains. However, deep neural networks are very resource-intensive in terms of energy …

A survey on efficient convolutional neural networks and hardware acceleration

D Ghimire, D Kil, S Kim - Electronics, 2022 - mdpi.com
Over the past decade, deep-learning-based representations have demonstrated remarkable
performance in academia and industry. The learning capability of convolutional neural …

Knowledge distillation: A survey

J Gou, B Yu, SJ Maybank, D Tao - International Journal of Computer Vision, 2021 - Springer
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …

Knowledge distillation and student-teacher learning for visual intelligence: A review and new outlooks

L Wang, KJ Yoon - IEEE transactions on pattern analysis and …, 2021 - ieeexplore.ieee.org
Deep neural models, in recent years, have been successful in almost every field, even
solving the most complex problem statements. However, these models are huge in size with …

Curriculum temperature for knowledge distillation

Z Li, X Li, L Yang, B Zhao, R Song, L Luo, J Li… - Proceedings of the …, 2023 - ojs.aaai.org
Most existing distillation methods ignore the flexible role of the temperature in the loss
function and fix it as a hyper-parameter that can be decided by an inefficient grid search. In …

Self-distillation: Towards efficient and compact neural networks

L Zhang, C Bao, K Ma - IEEE Transactions on Pattern Analysis …, 2021 - ieeexplore.ieee.org
Remarkable achievements have been obtained by deep neural networks in the last several
years. However, the breakthrough in neural networks accuracy is always accompanied by …

A survey on green deep learning

J Xu, W Zhou, Z Fu, H Zhou, L Li - arxiv preprint arxiv:2111.05193, 2021 - arxiv.org
In recent years, larger and deeper models are springing up and continuously pushing state-
of-the-art (SOTA) results across various fields like natural language processing (NLP) and …

Spot-adaptive knowledge distillation

J Song, Y Chen, J Ye, M Song - IEEE Transactions on Image …, 2022 - ieeexplore.ieee.org
Knowledge distillation (KD) has become a well established paradigm for compressing deep
neural networks. The typical way of conducting knowledge distillation is to train the student …

TransKD: Transformer knowledge distillation for efficient semantic segmentation

R Liu, K Yang, A Roitberg, J Zhang, K Peng… - arxiv preprint arxiv …, 2022 - arxiv.org
Semantic segmentation benchmarks in the realm of autonomous driving are dominated by
large pre-trained transformers, yet their widespread adoption is impeded by substantial …

Ernie 3.0 titan: Exploring larger-scale knowledge enhanced pre-training for language understanding and generation

S Wang, Y Sun, Y **ang, Z Wu, S Ding, W Gong… - arxiv preprint arxiv …, 2021 - arxiv.org
Pre-trained language models have achieved state-of-the-art results in various Natural
Language Processing (NLP) tasks. GPT-3 has shown that scaling up pre-trained language …