Prior gradient mask guided pruning-aware fine-tuning

L Cai, Z An, C Yang, Y Yan, Y Xu - … of the AAAI Conference on Artificial …, 2022 - ojs.aaai.org
Abstract We proposed a Prior Gradient Mask Guided Pruning-aware Fine-Tuning (PGMPF)
framework to accelerate deep Convolutional Neural Networks (CNNs). In detail, the …

CoAxNN: Optimizing on-device deep learning with conditional approximate neural networks

G Li, X Ma, Q Yu, L Liu, H Liu, X Wang - Journal of Systems Architecture, 2023 - Elsevier
While deep neural networks have achieved superior performance in a variety of intelligent
applications, the increasing computational complexity makes them difficult to be deployed …

Accelerating deep neural network filter pruning with mask-aware convolutional computations on modern CPUs

X Ma, G Li, L Liu, H Liu, X Wang - Neurocomputing, 2022 - Elsevier
Filter pruning, a representative model compression technique, has been widely used to
compress and accelerate sophisticated deep neural networks on resource-constrained …

Pruning-and-distillation: One-stage joint compression framework for CNNs via clustering

T Niu, Y Teng, L **, P Zou, Y Liu - Image and Vision Computing, 2023 - Elsevier
Network pruning and knowledge distillation, as two effective network compression
techniques, have drawn extensive attention due to their success in reducing model …

A differentiable framework for end-to-end learning of hybrid structured compression

M Eo, S Kang, W Rhee - arxiv preprint arxiv:2309.13077, 2023 - arxiv.org
Filter pruning and low-rank decomposition are two of the foundational techniques for
structured compression. Although recent efforts have explored hybrid approaches aiming to …

Enhancing accuracy of compressed Convolutional Neural Networks through a transfer teacher and reinforcement guided training curriculum

A Jayasimhan, P Pabitha - Knowledge-Based Systems, 2024 - Elsevier
Abstract Model compression techniques, such as network pruning, quantization and
knowledge distillation, are essential for deploying large Convolutional Neural Networks …

Asymptotic soft cluster pruning for deep neural networks

T Niu, Y Teng, P Zou - arxiv preprint arxiv:2206.08186, 2022 - arxiv.org
Filter pruning method introduces structural sparsity by removing selected filters and is thus
particularly effective for reducing complexity. Previous works empirically prune networks …

Soft Hybrid Filter Pruning using a Dual Ranking Approach

PY Chen, JC Yang, SD Wang - … on Trust, Security and Privacy in …, 2023 - ieeexplore.ieee.org
Conventional pruning techniques typically focus on evaluating a single structure in the
network, such as the convolutional layer or batch normalization layer, to identify pruning …

Lossless Filter Pruning via Adaptive Clustering for Convolutional Neural Networks

T Niu, Y Teng, P Zou, Y Liu - 2023 - openreview.net
The filter pruning method introduces structural sparsity by removing selected filters and is
thus particularly effective for reducing complexity. However, previous works face two …

LESS: LEARNING TO SELECT A STRUCTURED ARCHITECTURE OVER FILTER PRUNING AND LOW-RANK DECOMPOSITION

M Eo, S Kang, W Rhee - 5th Workshop on practical ML for limited/low … - openreview.net
Designing a deep neural network (DNN) for efficient operation in low-resource environments
necessitates strategic application of compression techniques. Filter pruning and low-rank …