Spvit: Enabling faster vision transformers via latency-aware soft token pruning

Z Kong, P Dong, X Ma, X Meng, W Niu, M Sun… - European conference on …, 2022 - Springer
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …

Chex: Channel exploration for cnn model compression

Z Hou, M Qin, F Sun, X Ma, K Yuan… - Proceedings of the …, 2022 - openaccess.thecvf.com
Channel pruning has been broadly recognized as an effective technique to reduce the
computation and memory cost of deep convolutional neural networks. However …

Multidimensional pruning and its extension: A unified framework for model compression

J Guo, D Xu, W Ouyang - IEEE Transactions on Neural …, 2023 - ieeexplore.ieee.org
Observing that the existing model compression approaches only focus on reducing the
redundancies in convolutional neural networks (CNNs) along one particular dimension (eg …

Deep compression for efficient and accelerated over-the-air federated learning

FMA Khan, H Abou-Zeid… - IEEE Internet of Things …, 2024 - ieeexplore.ieee.org
Over-the-air federated learning (OTA-FL) is a distributed machine learning technique where
multiple devices collaboratively train a shared model without sharing their raw data with a …

Compiler-aware neural architecture search for on-mobile real-time super-resolution

Y Wu, Y Gong, P Zhao, Y Li, Z Zhan, W Niu… - … on Computer Vision, 2022 - Springer
Deep learning-based super-resolution (SR) has gained tremendous popularity in recent
years because of its high image quality performance and wide application scenarios …

Effective model sparsification by scheduled grow-and-prune methods

X Ma, M Qin, F Sun, Z Hou, K Yuan, Y Xu… - arxiv preprint arxiv …, 2021 - arxiv.org
Deep neural networks (DNNs) are effective in solving many real-world problems. Larger
DNN models usually exhibit better quality (eg, accuracy) but their excessive computation …

Aedfl: efficient asynchronous decentralized federated learning with heterogeneous devices

J Liu, T Che, Y Zhou, R **, H Dai, D Dou… - Proceedings of the 2024 …, 2024 - SIAM
Federated Learning (FL) has achieved significant achievements recently, enabling
collaborative model training on distributed data over edge devices. Iterative gradient or …

An automatic and efficient BERT pruning for edge AI systems

S Huang, N Liu, Y Liang, H Peng, H Li… - … on Quality Electronic …, 2022 - ieeexplore.ieee.org
With the yearning for deep learning democratization, there are increasing demands to
implement Transformer-based natural language processing (NLP) models on resource …

Surface-related multiple attenuation based on a self-supervised deep neural network with local wavefield characteristics

K Wang, T Hu, B Zhao, S Wang - Geophysics, 2023 - library.seg.org
Multiple suppression is a very important step in seismic data processing. To suppress
surface-related multiples, we develop a self-supervised deep neural network method based …

Enabling fast deep learning on tiny energy-harvesting IoT devices

S Islam, J Deng, S Zhou, C Pan… - … Design, Automation & …, 2022 - ieeexplore.ieee.org
Energy harvesting (EH) IoT devices that operate intermittently without batteries, coupled with
advances in deep neural networks (DNNs), have opened up new opportunities for en-abling …