Model compression and acceleration for deep neural networks: The principles, progress, and challenges

Y Cheng, D Wang, P Zhou… - IEEE Signal Processing …, 2018 - ieeexplore.ieee.org
In recent years, deep neural networks (DNNs) have received increased attention, have been
applied to different applications, and achieved dramatic accuracy improvements in many …

Deep learning for generic object detection: A survey

L Liu, W Ouyang, X Wang, P Fieguth, J Chen… - International journal of …, 2020 - Springer
Object detection, one of the most fundamental and challenging problems in computer vision,
seeks to locate object instances from a large number of predefined categories in natural …

Exploring deep learning-based architecture, strategies, applications and current trends in generic object detection: A comprehensive review

L Aziz, MSBH Salam, UU Sheikh, S Ayub - Ieee Access, 2020 - ieeexplore.ieee.org
Object detection is a fundamental but challenging issue in the field of generic image
analysis; it plays an important role in a wide range of applications and has been receiving …

Channel pruning for accelerating very deep neural networks

Y He, X Zhang, J Sun - Proceedings of the IEEE …, 2017 - openaccess.thecvf.com
In this paper, we introduce a new channel pruning method to accelerate very deep
convolutional neural networks. Given a trained CNN model, we propose an iterative two …

Rethinking the value of network pruning

Z Liu, M Sun, T Zhou, G Huang, T Darrell - arxiv preprint arxiv:1810.05270, 2018 - arxiv.org
Network pruning is widely used for reducing the heavy inference cost of deep models in low-
resource settings. A typical pruning algorithm is a three-stage pipeline, ie, training (a large …

Learning structured sparsity in deep neural networks

W Wen, C Wu, Y Wang, Y Chen… - Advances in neural …, 2016 - proceedings.neurips.cc
High demand for computation resources severely hinders deployment of large-scale Deep
Neural Networks (DNN) in resource constrained devices. In this work, we propose a …

Pruning convolutional neural networks for resource efficient inference

P Molchanov, S Tyree, T Karras, T Aila… - arxiv preprint arxiv …, 2016 - arxiv.org
We propose a new formulation for pruning convolutional kernels in neural networks to
enable efficient inference. We interleave greedy criteria-based pruning with fine-tuning by …

To prune, or not to prune: exploring the efficacy of pruning for model compression

M Zhu, S Gupta - arxiv preprint arxiv:1710.01878, 2017 - arxiv.org
Model pruning seeks to induce sparsity in a deep neural network's various connection
matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent …

A survey of model compression and acceleration for deep neural networks

Y Cheng, D Wang, P Zhou, T Zhang - arxiv preprint arxiv:1710.09282, 2017 - arxiv.org
Deep neural networks (DNNs) have recently achieved great success in many visual
recognition tasks. However, existing deep neural network models are computationally …

Lifelong learning with dynamically expandable networks

J Yoon, E Yang, J Lee, SJ Hwang - arxiv preprint arxiv:1708.01547, 2017 - arxiv.org
We propose a novel deep network architecture for lifelong learning which we refer to as
Dynamically Expandable Network (DEN), that can dynamically decide its network capacity …