A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …
massive model sizes that require significant computational and storage resources. To …
Semantic communications for future internet: Fundamentals, applications, and challenges
With the increasing demand for intelligent services, the sixth-generation (6G) wireless
networks will shift from a traditional architecture that focuses solely on a high transmission …
networks will shift from a traditional architecture that focuses solely on a high transmission …
Depgraph: Towards any structural pruning
Structural pruning enables model acceleration by removing structurally-grouped parameters
from neural networks. However, the parameter-grou** patterns vary widely across …
from neural networks. However, the parameter-grou** patterns vary widely across …
A simple and effective pruning approach for large language models
As their size increases, Large Languages Models (LLMs) are natural candidates for network
pruning methods: approaches that drop a subset of network weights while striving to …
pruning methods: approaches that drop a subset of network weights while striving to …
Patch diffusion: Faster and more data-efficient training of diffusion models
Diffusion models are powerful, but they require a lot of time and data to train. We propose
Patch Diffusion, a generic patch-wise training framework, to significantly reduce the training …
Patch Diffusion, a generic patch-wise training framework, to significantly reduce the training …
Snapfusion: Text-to-image diffusion model on mobile devices within two seconds
Text-to-image diffusion models can create stunning images from natural language
descriptions that rival the work of professional artists and photographers. However, these …
descriptions that rival the work of professional artists and photographers. However, these …
On-device training under 256kb memory
On-device training enables the model to adapt to new data collected from the sensors by
fine-tuning a pre-trained model. Users can benefit from customized AI models without having …
fine-tuning a pre-trained model. Users can benefit from customized AI models without having …
Scconv: Spatial and channel reconstruction convolution for feature redundancy
J Li, Y Wen, L He - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Abstract Convolutional Neural Networks (CNNs) have achieved remarkable performance in
various computer vision tasks but this comes at the cost of tremendous computational …
various computer vision tasks but this comes at the cost of tremendous computational …
Sheared llama: Accelerating language model pre-training via structured pruning
The popularity of LLaMA (Touvron et al., 2023a; b) and other recently emerged moderate-
sized large language models (LLMs) highlights the potential of building smaller yet powerful …
sized large language models (LLMs) highlights the potential of building smaller yet powerful …
Efficientvit: Lightweight multi-scale attention for high-resolution dense prediction
High-resolution dense prediction enables many appealing real-world applications, such as
computational photography, autonomous driving, etc. However, the vast computational cost …
computational photography, autonomous driving, etc. However, the vast computational cost …