A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Chex: Channel exploration for cnn model compression

Z Hou, M Qin, F Sun, X Ma, K Yuan… - Proceedings of the …, 2022 - openaccess.thecvf.com
Channel pruning has been broadly recognized as an effective technique to reduce the
computation and memory cost of deep convolutional neural networks. However …

Sparcl: Sparse continual learning on the edge

Z Wang, Z Zhan, Y Gong, G Yuan… - Advances in …, 2022 - proceedings.neurips.cc
Existing work in continual learning (CL) focuses on mitigating catastrophic forgetting, ie,
model performance deterioration on past tasks when learning a new task. However, the …

Neural pruning via growing regularization

H Wang, C Qin, Y Zhang, Y Fu - arxiv preprint arxiv:2012.09243, 2020 - arxiv.org
Regularization has long been utilized to learn sparsity in deep neural network pruning.
However, its role is mainly explored in the small penalty strength regime. In this work, we …

Mest: Accurate and fast memory-economic sparse training framework on the edge

G Yuan, X Ma, W Niu, Z Li, Z Kong… - Advances in …, 2021 - proceedings.neurips.cc
Recently, a new trend of exploring sparsity for accelerating neural network training has
emerged, embracing the paradigm of training on the edge. This paper proposes a novel …

Yolobile: Real-time object detection on mobile devices via compression-compilation co-design

Y Cai, H Li, G Yuan, W Niu, Y Li, X Tang… - Proceedings of the …, 2021 - ojs.aaai.org
The rapid development and wide utilization of object detection techniques have aroused
attention on both accuracy and speed of object detectors. However, the current state-of-the …

Iterative clustering pruning for convolutional neural networks

J Chang, Y Lu, P Xue, Y Xu, Z Wei - Knowledge-Based Systems, 2023 - Elsevier
Convolutional neural networks (CNNs) have shown excellent performance in numerous
computer vision tasks. However, the high computational and memory demands in computer …

Deephoyer: Learning sparser neural network with differentiable scale-invariant sparsity measures

H Yang, W Wen, H Li - arxiv preprint arxiv:1908.09979, 2019 - arxiv.org
In seeking for sparse and efficient neural network models, many previous works investigated
on enforcing L1 or L0 regularizers to encourage weight sparsity during training. The L0 …

Sparse adversarial attack via perturbation factorization

Y Fan, B Wu, T Li, Y Zhang, M Li, Z Li… - Computer Vision–ECCV …, 2020 - Springer
This work studies the sparse adversarial attack, which aims to generate adversarial
perturbations onto partial positions of one benign image, such that the perturbed image is …

Trainability preserving neural pruning

H Wang, Y Fu - arxiv preprint arxiv:2207.12534, 2022 - arxiv.org
Many recent works have shown trainability plays a central role in neural network pruning--
unattended broken trainability can lead to severe under-performance and unintentionally …