Prior gradient mask guided pruning-aware fine-tuning
Abstract We proposed a Prior Gradient Mask Guided Pruning-aware Fine-Tuning (PGMPF)
framework to accelerate deep Convolutional Neural Networks (CNNs). In detail, the …
framework to accelerate deep Convolutional Neural Networks (CNNs). In detail, the …
CoAxNN: Optimizing on-device deep learning with conditional approximate neural networks
While deep neural networks have achieved superior performance in a variety of intelligent
applications, the increasing computational complexity makes them difficult to be deployed …
applications, the increasing computational complexity makes them difficult to be deployed …
Accelerating deep neural network filter pruning with mask-aware convolutional computations on modern CPUs
Filter pruning, a representative model compression technique, has been widely used to
compress and accelerate sophisticated deep neural networks on resource-constrained …
compress and accelerate sophisticated deep neural networks on resource-constrained …
Pruning-and-distillation: One-stage joint compression framework for CNNs via clustering
Network pruning and knowledge distillation, as two effective network compression
techniques, have drawn extensive attention due to their success in reducing model …
techniques, have drawn extensive attention due to their success in reducing model …
A differentiable framework for end-to-end learning of hybrid structured compression
Filter pruning and low-rank decomposition are two of the foundational techniques for
structured compression. Although recent efforts have explored hybrid approaches aiming to …
structured compression. Although recent efforts have explored hybrid approaches aiming to …
Enhancing accuracy of compressed Convolutional Neural Networks through a transfer teacher and reinforcement guided training curriculum
Abstract Model compression techniques, such as network pruning, quantization and
knowledge distillation, are essential for deploying large Convolutional Neural Networks …
knowledge distillation, are essential for deploying large Convolutional Neural Networks …
Asymptotic soft cluster pruning for deep neural networks
T Niu, Y Teng, P Zou - arxiv preprint arxiv:2206.08186, 2022 - arxiv.org
Filter pruning method introduces structural sparsity by removing selected filters and is thus
particularly effective for reducing complexity. Previous works empirically prune networks …
particularly effective for reducing complexity. Previous works empirically prune networks …
Soft Hybrid Filter Pruning using a Dual Ranking Approach
PY Chen, JC Yang, SD Wang - … on Trust, Security and Privacy in …, 2023 - ieeexplore.ieee.org
Conventional pruning techniques typically focus on evaluating a single structure in the
network, such as the convolutional layer or batch normalization layer, to identify pruning …
network, such as the convolutional layer or batch normalization layer, to identify pruning …
Lossless Filter Pruning via Adaptive Clustering for Convolutional Neural Networks
T Niu, Y Teng, P Zou, Y Liu - 2023 - openreview.net
The filter pruning method introduces structural sparsity by removing selected filters and is
thus particularly effective for reducing complexity. However, previous works face two …
thus particularly effective for reducing complexity. However, previous works face two …
LESS: LEARNING TO SELECT A STRUCTURED ARCHITECTURE OVER FILTER PRUNING AND LOW-RANK DECOMPOSITION
Designing a deep neural network (DNN) for efficient operation in low-resource environments
necessitates strategic application of compression techniques. Filter pruning and low-rank …
necessitates strategic application of compression techniques. Filter pruning and low-rank …