A survey on federated learning for resource-constrained IoT devices

A Imteaj, U Thakker, S Wang, J Li… - IEEE Internet of Things …, 2021 - ieeexplore.ieee.org
Federated learning (FL) is a distributed machine learning strategy that generates a global
model by learning from multiple decentralized edge clients. FL enables on-device training …

Efficient acceleration of deep learning inference on resource-constrained edge devices: A review

MMH Shuvo, SK Islam, J Cheng… - Proceedings of the …, 2022 - ieeexplore.ieee.org
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …

Dynamic neural networks: A survey

Y Han, G Huang, S Song, L Yang… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …

Rethinking the value of network pruning

Z Liu, M Sun, T Zhou, G Huang, T Darrell - arxiv preprint arxiv:1810.05270, 2018 - arxiv.org
Network pruning is widely used for reducing the heavy inference cost of deep models in low-
resource settings. A typical pruning algorithm is a three-stage pipeline, ie, training (a large …

Dynamic convolution: Attention over convolution kernels

Y Chen, X Dai, M Liu, D Chen… - Proceedings of the …, 2020 - openaccess.thecvf.com
Light-weight convolutional neural networks (CNNs) suffer performance degradation as their
low computational budgets constrain both the depth (number of convolution layers) and the …

Lite-hrnet: A lightweight high-resolution network

C Yu, B **ao, C Gao, L Yuan, L Zhang… - Proceedings of the …, 2021 - openaccess.thecvf.com
We present an efficient high-resolution network, Lite-HRNet, for human pose estimation. We
start by simply applying the efficient shuffle block in ShuffleNet to HRNet (high-resolution …

Be your own teacher: Improve the performance of convolutional neural networks via self distillation

L Zhang, J Song, A Gao, J Chen… - Proceedings of the …, 2019 - openaccess.thecvf.com
Convolutional neural networks have been widely deployed in various application scenarios.
In order to extend the applications' boundaries to some accuracy-crucial domains …

A-vit: Adaptive tokens for efficient vision transformer

H Yin, A Vahdat, JM Alvarez, A Mallya… - Proceedings of the …, 2022 - openaccess.thecvf.com
We introduce A-ViT, a method that adaptively adjusts the inference cost of vision transformer
ViT for images of different complexity. A-ViT achieves this by automatically reducing the …

Not all patches are what you need: Expediting vision transformers via token reorganizations

Y Liang, C Ge, Z Tong, Y Song, J Wang… - arxiv preprint arxiv …, 2022 - arxiv.org
Vision Transformers (ViTs) take all the image patches as tokens and construct multi-head
self-attention (MHSA) among them. Complete leverage of these image tokens brings …

Reducing transformer depth on demand with structured dropout

A Fan, E Grave, A Joulin - arxiv preprint arxiv:1909.11556, 2019 - arxiv.org
Overparameterized transformer networks have obtained state of the art results in various
natural language processing tasks, such as machine translation, language modeling, and …