A survey on efficient convolutional neural networks and hardware acceleration

D Ghimire, D Kil, S Kim - Electronics, 2022 - mdpi.com
Over the past decade, deep-learning-based representations have demonstrated remarkable
performance in academia and industry. The learning capability of convolutional neural …

A review of convolutional neural network architectures and their optimizations

S Cong, Y Zhou - Artificial Intelligence Review, 2023 - Springer
The research advances concerning the typical architectures of convolutional neural
networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this …

Blueprint separable residual network for efficient image super-resolution

Z Li, Y Liu, X Chen, H Cai, J Gu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Recent advances in single image super-resolution (SISR) have achieved extraordinary
performance, but the computational cost is too heavy to apply in edge devices. To alleviate …

Dynamic neural networks: A survey

Y Han, G Huang, S Song, L Yang… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Ganspace: Discovering interpretable gan controls

E Härkönen, A Hertzmann… - Advances in neural …, 2020 - proceedings.neurips.cc
This paper describes a simple technique to analyze Generative Adversarial Networks
(GANs) and create interpretable controls for image synthesis, such as change of viewpoint …

Hrank: Filter pruning using high-rank feature map

M Lin, R Ji, Y Wang, Y Zhang… - Proceedings of the …, 2020 - openaccess.thecvf.com
Neural network pruning offers a promising prospect to facilitate deploying deep neural
networks on resource-limited devices. However, existing methods are still challenged by the …

Cross-layer distillation with semantic calibration

D Chen, JP Mei, Y Zhang, C Wang, Z Wang… - Proceedings of the …, 2021 - ojs.aaai.org
Recently proposed knowledge distillation approaches based on feature-map transfer
validate that intermediate layers of a teacher model can serve as effective targets for training …

Model compression and hardware acceleration for neural networks: A comprehensive survey

L Deng, G Li, S Han, L Shi, Y **e - Proceedings of the IEEE, 2020 - ieeexplore.ieee.org
Domain-specific hardware is becoming a promising topic in the backdrop of improvement
slow down for general-purpose processors due to the foreseeable end of Moore's Law …

Patient knowledge distillation for bert model compression

S Sun, Y Cheng, Z Gan, J Liu - arxiv preprint arxiv:1908.09355, 2019 - arxiv.org
Pre-trained language models such as BERT have proven to be highly effective for natural
language processing (NLP) tasks. However, the high demand for computing resources in …