A survey on federated learning for resource-constrained IoT devices
Federated learning (FL) is a distributed machine learning strategy that generates a global
model by learning from multiple decentralized edge clients. FL enables on-device training …
model by learning from multiple decentralized edge clients. FL enables on-device training …
Efficient acceleration of deep learning inference on resource-constrained edge devices: A review
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …
in breakthroughs in many areas. However, deploying these highly accurate models for data …
Dynamic neural networks: A survey
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …
models which have fixed computational graphs and parameters at the inference stage …
Rethinking the value of network pruning
Network pruning is widely used for reducing the heavy inference cost of deep models in low-
resource settings. A typical pruning algorithm is a three-stage pipeline, ie, training (a large …
resource settings. A typical pruning algorithm is a three-stage pipeline, ie, training (a large …
Dynamic convolution: Attention over convolution kernels
Light-weight convolutional neural networks (CNNs) suffer performance degradation as their
low computational budgets constrain both the depth (number of convolution layers) and the …
low computational budgets constrain both the depth (number of convolution layers) and the …
Lite-hrnet: A lightweight high-resolution network
We present an efficient high-resolution network, Lite-HRNet, for human pose estimation. We
start by simply applying the efficient shuffle block in ShuffleNet to HRNet (high-resolution …
start by simply applying the efficient shuffle block in ShuffleNet to HRNet (high-resolution …
Be your own teacher: Improve the performance of convolutional neural networks via self distillation
L Zhang, J Song, A Gao, J Chen… - Proceedings of the …, 2019 - openaccess.thecvf.com
Convolutional neural networks have been widely deployed in various application scenarios.
In order to extend the applications' boundaries to some accuracy-crucial domains …
In order to extend the applications' boundaries to some accuracy-crucial domains …
A-vit: Adaptive tokens for efficient vision transformer
We introduce A-ViT, a method that adaptively adjusts the inference cost of vision transformer
ViT for images of different complexity. A-ViT achieves this by automatically reducing the …
ViT for images of different complexity. A-ViT achieves this by automatically reducing the …
Not all patches are what you need: Expediting vision transformers via token reorganizations
Vision Transformers (ViTs) take all the image patches as tokens and construct multi-head
self-attention (MHSA) among them. Complete leverage of these image tokens brings …
self-attention (MHSA) among them. Complete leverage of these image tokens brings …
Reducing transformer depth on demand with structured dropout
Overparameterized transformer networks have obtained state of the art results in various
natural language processing tasks, such as machine translation, language modeling, and …
natural language processing tasks, such as machine translation, language modeling, and …