A survey of methods for low-power deep learning and computer vision

A Goel, C Tung, YH Lu… - 2020 IEEE 6th World …, 2020 - ieeexplore.ieee.org
Deep neural networks (DNNs) are successful in many computer vision tasks. However, the
most accurate DNNs require millions of parameters and operations, making them energy …

Quantization and deployment of deep neural networks on microcontrollers

PE Novac, G Boukli Hacene, A Pegatoquet… - Sensors, 2021 - mdpi.com
Embedding Artificial Intelligence onto low-power devices is a challenging task that has been
partly overcome with recent advances in machine learning and hardware design. Presently …

Autopruner: An end-to-end trainable filter pruning method for efficient deep model inference

JH Luo, J Wu - Pattern Recognition, 2020 - Elsevier
Channel pruning is an important method to speed up CNN model's inference. Previous filter
pruning algorithms regard importance evaluation and model fine-tuning as two independent …

Graph neural networks: Architectures, stability, and transferability

L Ruiz, F Gama, A Ribeiro - Proceedings of the IEEE, 2021 - ieeexplore.ieee.org
Graph neural networks (GNNs) are information processing architectures for signals
supported on graphs. They are presented here as generalizations of convolutional neural …

Rubiksnet: Learnable 3d-shift for efficient video action recognition

L Fan, S Buch, G Wang, R Cao, Y Zhu… - … on Computer Vision, 2020 - Springer
Video action recognition is a complex task dependent on modeling spatial and temporal
context. Standard approaches rely on 2D or 3D convolutions to process such context …

Complexity-driven model compression for resource-constrained deep learning on edge

M Zawish, S Davy, L Abraham - IEEE Transactions on Artificial …, 2024 - ieeexplore.ieee.org
Recent advances in artificial intelligence (AI) on the Internet of Things (IoT) devices have
realized edge AI in several applications by enabling low latency and energy efficiency …

Deep geometric knowledge distillation with graphs

C Lassance, M Bontonou, GB Hacene… - ICASSP 2020-2020 …, 2020 - ieeexplore.ieee.org
In most cases deep learning architectures are trained disregarding the amount of operations
and energy consumption. However, some applications, like embedded systems, can be …

Rethinking weight decay for efficient neural network pruning

H Tessier, V Gripon, M Léonardon, M Arzel… - Journal of …, 2022 - mdpi.com
Introduced in the late 1980s for generalization purposes, pruning has now become a staple
for compressing deep neural networks. Despite many innovations in recent decades …

EPSViTs: A hybrid architecture for image classification based on parameter-shared multi-head self-attention

H Liao, X Li, X Qin, W Wang, G He, H Huang… - Image and Vision …, 2024 - Elsevier
Vision transformers have been successfully applied to image recognition tasks due to their
ability to capture long-range dependencies within an image. However, they still suffer from …

Quantized guided pruning for efficient hardware implementations of deep neural networks

GB Hacene, V Gripon, M Arzel… - 2020 18th IEEE …, 2020 - ieeexplore.ieee.org
Deep Neural Networks (DNNs) in general and Convolutional Neural Networks (CNNs) in
particular are state-of-the-art in numerous computer vision tasks such as object classification …