A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …
massive model sizes that require significant computational and storage resources. To …
Recent advances on neural network pruning at initialization
Neural network pruning typically removes connections or neurons from a pretrained
converged model; while a new pruning paradigm, pruning at initialization (PaI), attempts to …
converged model; while a new pruning paradigm, pruning at initialization (PaI), attempts to …
Exploring lottery ticket hypothesis in spiking neural networks
Abstract Spiking Neural Networks (SNNs) have recently emerged as a new generation of
low-power deep neural networks, which is suitable to be implemented on low-power …
low-power deep neural networks, which is suitable to be implemented on low-power …
[PDF][PDF] Searching lottery tickets in graph neural networks: A dual perspective
Graph Neural Networks (GNNs) have shown great promise in various graph learning tasks.
However, the computational overheads of fitting GNNs to large-scale graphs grow rapidly …
However, the computational overheads of fitting GNNs to large-scale graphs grow rapidly …
The snowflake hypothesis: Training deep GNN with one node one receptive field
Despite Graph Neural Networks demonstrating considerable promise in graph
representation learning tasks, GNNs predominantly face significant issues with over-fitting …
representation learning tasks, GNNs predominantly face significant issues with over-fitting …
Parameter-efficient masking networks
A deeper network structure generally handles more complicated non-linearity and performs
more competitively. Nowadays, advanced network designs often contain a large number of …
more competitively. Nowadays, advanced network designs often contain a large number of …
Dimensionality reduced training by pruning and freezing parts of a deep neural network: a survey
State-of-the-art deep learning models have a parameter count that reaches into the billions.
Training, storing and transferring such models is energy and time consuming, thus costly. A …
Training, storing and transferring such models is energy and time consuming, thus costly. A …
Enhanced sparsification via stimulative training
Sparsification-based pruning has been an important category in model compression.
Existing methods commonly set sparsity-inducing penalty terms to suppress the importance …
Existing methods commonly set sparsity-inducing penalty terms to suppress the importance …
Learn from Downstream and Be Yourself in Multimodal Large Language Model Fine-Tuning
Multimodal Large Language Model (MLLM) have demonstrated strong generalization
capabilities across diverse distributions and tasks, largely due to extensive pre-training …
capabilities across diverse distributions and tasks, largely due to extensive pre-training …
Distributionally robust ensemble of lottery tickets towards calibrated sparse network training
The recently developed sparse network training methods, such as Lottery Ticket Hypothesis
(LTH) and its variants, have shown impressive learning capacity by finding sparse sub …
(LTH) and its variants, have shown impressive learning capacity by finding sparse sub …