Spvit: Enabling faster vision transformers via latency-aware soft token pruning
Abstract Recently, Vision Transformer (ViT) has continuously established new milestones in
the computer vision field, while the high computation and memory cost makes its …
the computer vision field, while the high computation and memory cost makes its …
Chex: Channel exploration for cnn model compression
Channel pruning has been broadly recognized as an effective technique to reduce the
computation and memory cost of deep convolutional neural networks. However …
computation and memory cost of deep convolutional neural networks. However …
Multidimensional pruning and its extension: A unified framework for model compression
Observing that the existing model compression approaches only focus on reducing the
redundancies in convolutional neural networks (CNNs) along one particular dimension (eg …
redundancies in convolutional neural networks (CNNs) along one particular dimension (eg …
Deep compression for efficient and accelerated over-the-air federated learning
FMA Khan, H Abou-Zeid… - IEEE Internet of Things …, 2024 - ieeexplore.ieee.org
Over-the-air federated learning (OTA-FL) is a distributed machine learning technique where
multiple devices collaboratively train a shared model without sharing their raw data with a …
multiple devices collaboratively train a shared model without sharing their raw data with a …
Compiler-aware neural architecture search for on-mobile real-time super-resolution
Deep learning-based super-resolution (SR) has gained tremendous popularity in recent
years because of its high image quality performance and wide application scenarios …
years because of its high image quality performance and wide application scenarios …
Effective model sparsification by scheduled grow-and-prune methods
Deep neural networks (DNNs) are effective in solving many real-world problems. Larger
DNN models usually exhibit better quality (eg, accuracy) but their excessive computation …
DNN models usually exhibit better quality (eg, accuracy) but their excessive computation …
Aedfl: efficient asynchronous decentralized federated learning with heterogeneous devices
Federated Learning (FL) has achieved significant achievements recently, enabling
collaborative model training on distributed data over edge devices. Iterative gradient or …
collaborative model training on distributed data over edge devices. Iterative gradient or …
An automatic and efficient BERT pruning for edge AI systems
With the yearning for deep learning democratization, there are increasing demands to
implement Transformer-based natural language processing (NLP) models on resource …
implement Transformer-based natural language processing (NLP) models on resource …
Surface-related multiple attenuation based on a self-supervised deep neural network with local wavefield characteristics
K Wang, T Hu, B Zhao, S Wang - Geophysics, 2023 - library.seg.org
Multiple suppression is a very important step in seismic data processing. To suppress
surface-related multiples, we develop a self-supervised deep neural network method based …
surface-related multiples, we develop a self-supervised deep neural network method based …
Enabling fast deep learning on tiny energy-harvesting IoT devices
Energy harvesting (EH) IoT devices that operate intermittently without batteries, coupled with
advances in deep neural networks (DNNs), have opened up new opportunities for en-abling …
advances in deep neural networks (DNNs), have opened up new opportunities for en-abling …