A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Structured pruning for deep convolutional neural networks: A survey

Y He, L **ao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

On the opportunities of green computing: A survey

Y Zhou, X Lin, X Zhang, M Wang, G Jiang, H Lu… - arxiv preprint arxiv …, 2023 - arxiv.org
Artificial Intelligence (AI) has achieved significant advancements in technology and research
with the development over several decades, and is widely used in many areas including …

[HTML][HTML] Structured pruning adapters

L Hedegaard, A Alok, J Jose, A Iosifidis - Pattern Recognition, 2024 - Elsevier
Adapters are a parameter-efficient alternative to fine-tuning, which augment a frozen base
network to learn new tasks. Yet, the inference of the adapted model is often slower than the …

Singe: Sparsity via integrated gradients estimation of neuron relevance

E Yvinec, A Dapogny, M Cord… - Advances in Neural …, 2022 - proceedings.neurips.cc
The leap in performance in state-of-the-art computer vision methods is attributed to the
development of deep neural networks. However it often comes at a computational price …

A Multiply-And-Max/min Neuron Paradigm for Aggressively Prunable Deep Neural Networks

L Prono, P Bich, C Boretti, M Mangia… - … on Neural Networks …, 2025 - ieeexplore.ieee.org
The growing interest in the Internet of Things (IoT) and mobile artificial intelligence
applications is pushing the investigation on deep neural networks (DNNs) that can operate …

Compressing convolutional neural networks with hierarchical Tucker-2 decomposition

M Gabor, R Zdunek - Applied Soft Computing, 2023 - Elsevier
Convolutional neural networks (CNNs) play a crucial role and achieve top results in
computer vision tasks but at the cost of high computational cost and storage complexity. One …

A survey of lottery ticket hypothesis

B Liu, Z Zhang, P He, Z Wang, Y **ao, R Ye… - arxiv preprint arxiv …, 2024 - arxiv.org
The Lottery Ticket Hypothesis (LTH) states that a dense neural network model contains a
highly sparse subnetwork (ie, winning tickets) that can achieve even better performance …

A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts

MNR Chowdhury, M Wang, KE Maghraoui… - arxiv preprint arxiv …, 2024 - arxiv.org
The sparsely gated mixture of experts (MoE) architecture sends different inputs to different
subnetworks, ie, experts, through trainable routers. MoE reduces the training computation …

Pruning-and-distillation: One-stage joint compression framework for CNNs via clustering

T Niu, Y Teng, L **, P Zou, Y Liu - Image and Vision Computing, 2023 - Elsevier
Network pruning and knowledge distillation, as two effective network compression
techniques, have drawn extensive attention due to their success in reducing model …