Model compression and acceleration for deep neural networks: The principles, progress, and challenges
In recent years, deep neural networks (DNNs) have received increased attention, have been
applied to different applications, and achieved dramatic accuracy improvements in many …
applied to different applications, and achieved dramatic accuracy improvements in many …
Deep learning for generic object detection: A survey
Object detection, one of the most fundamental and challenging problems in computer vision,
seeks to locate object instances from a large number of predefined categories in natural …
seeks to locate object instances from a large number of predefined categories in natural …
Exploring deep learning-based architecture, strategies, applications and current trends in generic object detection: A comprehensive review
Object detection is a fundamental but challenging issue in the field of generic image
analysis; it plays an important role in a wide range of applications and has been receiving …
analysis; it plays an important role in a wide range of applications and has been receiving …
Channel pruning for accelerating very deep neural networks
In this paper, we introduce a new channel pruning method to accelerate very deep
convolutional neural networks. Given a trained CNN model, we propose an iterative two …
convolutional neural networks. Given a trained CNN model, we propose an iterative two …
Rethinking the value of network pruning
Network pruning is widely used for reducing the heavy inference cost of deep models in low-
resource settings. A typical pruning algorithm is a three-stage pipeline, ie, training (a large …
resource settings. A typical pruning algorithm is a three-stage pipeline, ie, training (a large …
Learning structured sparsity in deep neural networks
High demand for computation resources severely hinders deployment of large-scale Deep
Neural Networks (DNN) in resource constrained devices. In this work, we propose a …
Neural Networks (DNN) in resource constrained devices. In this work, we propose a …
Pruning convolutional neural networks for resource efficient inference
We propose a new formulation for pruning convolutional kernels in neural networks to
enable efficient inference. We interleave greedy criteria-based pruning with fine-tuning by …
enable efficient inference. We interleave greedy criteria-based pruning with fine-tuning by …
To prune, or not to prune: exploring the efficacy of pruning for model compression
M Zhu, S Gupta - arxiv preprint arxiv:1710.01878, 2017 - arxiv.org
Model pruning seeks to induce sparsity in a deep neural network's various connection
matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent …
matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent …
A survey of model compression and acceleration for deep neural networks
Deep neural networks (DNNs) have recently achieved great success in many visual
recognition tasks. However, existing deep neural network models are computationally …
recognition tasks. However, existing deep neural network models are computationally …
Lifelong learning with dynamically expandable networks
We propose a novel deep network architecture for lifelong learning which we refer to as
Dynamically Expandable Network (DEN), that can dynamically decide its network capacity …
Dynamically Expandable Network (DEN), that can dynamically decide its network capacity …