A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Recent advances on neural network pruning at initialization

H Wang, C Qin, Y Bai, Y Zhang, Y Fu - arxiv preprint arxiv:2103.06460, 2021 - arxiv.org
Neural network pruning typically removes connections or neurons from a pretrained
converged model; while a new pruning paradigm, pruning at initialization (PaI), attempts to …

Exploring lottery ticket hypothesis in spiking neural networks

Y Kim, Y Li, H Park, Y Venkatesha, R Yin… - European Conference on …, 2022 - Springer
Abstract Spiking Neural Networks (SNNs) have recently emerged as a new generation of
low-power deep neural networks, which is suitable to be implemented on low-power …

[PDF][PDF] Searching lottery tickets in graph neural networks: A dual perspective

K Wang, Y Liang, P Wang, X Wang, P Gu… - The Eleventh …, 2022 - openreview.net
Graph Neural Networks (GNNs) have shown great promise in various graph learning tasks.
However, the computational overheads of fitting GNNs to large-scale graphs grow rapidly …

The snowflake hypothesis: Training deep GNN with one node one receptive field

K Wang, G Li, S Wang, G Zhang, K Wang, Y You… - arxiv preprint arxiv …, 2023 - arxiv.org
Despite Graph Neural Networks demonstrating considerable promise in graph
representation learning tasks, GNNs predominantly face significant issues with over-fitting …

Parameter-efficient masking networks

Y Bai, H Wang, X Ma, Y Zhang… - Advances in Neural …, 2022 - proceedings.neurips.cc
A deeper network structure generally handles more complicated non-linearity and performs
more competitively. Nowadays, advanced network designs often contain a large number of …

Dimensionality reduced training by pruning and freezing parts of a deep neural network: a survey

P Wimmer, J Mehnert, AP Condurache - Artificial Intelligence Review, 2023 - Springer
State-of-the-art deep learning models have a parameter count that reaches into the billions.
Training, storing and transferring such models is energy and time consuming, thus costly. A …

Enhanced sparsification via stimulative training

S Tang, W Lin, H Ye, P Ye, C Yu, B Li… - European Conference on …, 2024 - Springer
Sparsification-based pruning has been an important category in model compression.
Existing methods commonly set sparsity-inducing penalty terms to suppress the importance …

Learn from Downstream and Be Yourself in Multimodal Large Language Model Fine-Tuning

W Huang, J Liang, Z Shi, D Zhu, G Wan, H Li… - arxiv preprint arxiv …, 2024 - arxiv.org
Multimodal Large Language Model (MLLM) have demonstrated strong generalization
capabilities across diverse distributions and tasks, largely due to extensive pre-training …

Distributionally robust ensemble of lottery tickets towards calibrated sparse network training

H Sapkota, D Wang, Z Tao… - Advances in Neural …, 2024 - proceedings.neurips.cc
The recently developed sparse network training methods, such as Lottery Ticket Hypothesis
(LTH) and its variants, have shown impressive learning capacity by finding sparse sub …