A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

A survey on federated learning for resource-constrained IoT devices

A Imteaj, U Thakker, S Wang, J Li… - IEEE Internet of Things …, 2021 - ieeexplore.ieee.org
Federated learning (FL) is a distributed machine learning strategy that generates a global
model by learning from multiple decentralized edge clients. FL enables on-device training …

Dynamic neural networks: A survey

Y Han, G Huang, S Song, L Yang… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …

Pruning and quantization for deep neural network acceleration: A survey

T Liang, J Glossner, L Wang, S Shi, X Zhang - Neurocomputing, 2021 - Elsevier
Deep neural networks have been applied in many applications exhibiting extraordinary
abilities in the field of computer vision. However, complex network architectures challenge …

Structured pruning for deep convolutional neural networks: A survey

Y He, L **ao - IEEE transactions on pattern analysis and …, 2023 - ieeexplore.ieee.org
The remarkable performance of deep Convolutional neural networks (CNNs) is generally
attributed to their deeper and wider architectures, which can come with significant …

Patch slimming for efficient vision transformers

Y Tang, K Han, Y Wang, C Xu, J Guo… - Proceedings of the …, 2022 - openaccess.thecvf.com
This paper studies the efficiency problem for visual transformers by excavating redundant
calculation in given networks. The recent transformer architecture has demonstrated its …

Dynamic slimmable network

C Li, G Wang, B Wang, X Liang… - Proceedings of the …, 2021 - openaccess.thecvf.com
Current dynamic networks and dynamic pruning methods have shown their promising
capability in reducing theoretical computation complexity. However, dynamic sparse …

Exploring sparsity in image super-resolution for efficient inference

L Wang, X Dong, Y Wang, X Ying… - Proceedings of the …, 2021 - openaccess.thecvf.com
Current CNN-based super-resolution (SR) methods process all locations equally with
computational resources being uniformly assigned in space. However, since missing details …

When the curious abandon honesty: Federated learning is not private

F Boenisch, A Dziedzic, R Schuster… - 2023 IEEE 8th …, 2023 - ieeexplore.ieee.org
In federated learning (FL), data does not leave personal devices when they are jointly
training a machine learning model. Instead, these devices share gradients, parameters, or …

Stable low-rank tensor decomposition for compression of convolutional neural network

AH Phan, K Sobolev, K Sozykin, D Ermilov… - Computer Vision–ECCV …, 2020 - Springer
Most state-of-the-art deep neural networks are overparameterized and exhibit a high
computational cost. A straightforward approach to this problem is to replace convolutional …