A heuristic exploration of retraining-free weight-sharing for CNN compression

E Dupuis, D Novo, I O'Connor… - 2022 27th Asia and …, 2022 - ieeexplore.ieee.org
The computational workload involved in Convolutional Neural Networks (CNNs) is typically
out of reach for low-power embedded devices. The scientific literature provides a large …

A new clustering-based technique for the acceleration of deep convolutional networks

EV Pikoulis, C Mavrokefalidis, S Nousias… - … Applications, Volume 3, 2022 - Springer
Deep learning and especially the use of Deep Neural Networks (DNNs) provide impressive
results in various regression and classification tasks. However, to achieve these results …

Ternary quantization: A survey

D Liu, X Liu - arxiv preprint arxiv:2303.01505, 2023 - arxiv.org
Inference time, model size, and accuracy are critical for deploying deep neural network
models. Numerous research efforts have been made to compress neural network models …

[HTML][HTML] Compressing deep networks by neuron agglomerative clustering

LN Wang, W Liu, X Liu, G Zhong, PP Roy, J Dong… - Sensors, 2020 - mdpi.com
In recent years, deep learning models have achieved remarkable successes in various
applications, such as pattern recognition, computer vision, and signal processing. However …

Second-order response transform attention network for image classification

J Zhang, J Wang, Q Sun, C Li, B Liu, Q Zhang… - IEEE …, 2019 - ieeexplore.ieee.org
Embedding second-order operations into deep convolutional neural networks (CNNs) has
recently shown impressive performance for a number of vision tasks. Specifically, the two …

Hyperspherical Loss-Aware Ternary Quantization

D Liu, X Liu - arxiv preprint arxiv:2212.12649, 2022 - arxiv.org
Most of the existing works use projection functions for ternary quantization in discrete space.
Scaling factors and thresholds are used in some cases to improve the model accuracy …

A New Clustering-Based Technique for the Acceleration of Deep

EV Pikoulis, C Mavrokefalidis… - … Applications, Volume 3, 2021 - books.google.com
Deep learning and especially the use of Deep Neural Networks (DNNs) provide impressive
results in various regression and classification tasks. However, to achieve these results …

[PDF][PDF] Weight-Sharing Methods for Retraining-Free CNN Compression

MD Pr - 2022 - bibliotheque.ec-lyon.fr
The outstanding performance achieved by Convolutional Neural Network (CNN) comes at
the cost of extremely high computational requirements, making them out of reach for most …

Méthode de partage de poids pour compresser des réseaux de neurones profonds sans réentraînement

E Dupuis - 2022 - theses.hal.science
Les performances exceptionnelles atteintes par les Convolutional Neural Network (CNN) se
font au prix d'exigences de calcul extrêmement élevées, les rendant hors de portée de la …