Artificial intelligence in the creative industries: a review
This paper reviews the current state of the art in artificial intelligence (AI) technologies and
applications in the context of the creative industries. A brief background of AI, and …
applications in the context of the creative industries. A brief background of AI, and …
A review of convolutional neural network architectures and their optimizations
The research advances concerning the typical architectures of convolutional neural
networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this …
networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this …
Brecq: Pushing the limit of post-training quantization by block reconstruction
We study the challenging task of neural network quantization without end-to-end retraining,
called Post-training Quantization (PTQ). PTQ usually requires a small subset of training data …
called Post-training Quantization (PTQ). PTQ usually requires a small subset of training data …
Training neural networks with fixed sparse masks
During typical gradient-based training of deep neural networks, all of the model's
parameters are updated at each iteration. Recent work has shown that it is possible to …
parameters are updated at each iteration. Recent work has shown that it is possible to …
Importance estimation for neural network pruning
Structural pruning of neural network parameters reduces computational, energy, and
memory transfer costs during inference. We propose a novel method that estimates the …
memory transfer costs during inference. We propose a novel method that estimates the …
Zero-cost proxies for lightweight nas
Neural Architecture Search (NAS) is quickly becoming the standard methodology to design
neural network models. However, NAS is typically compute-intensive because multiple …
neural network models. However, NAS is typically compute-intensive because multiple …
Movement pruning: Adaptive sparsity by fine-tuning
Magnitude pruning is a widely used strategy for reducing model size in pure supervised
learning; however, it is less effective in the transfer learning regime that has become …
learning; however, it is less effective in the transfer learning regime that has become …
Neural architecture search: Insights from 1000 papers
In the past decade, advances in deep learning have resulted in breakthroughs in a variety of
areas, including computer vision, natural language understanding, speech recognition, and …
areas, including computer vision, natural language understanding, speech recognition, and …
A fast post-training pruning framework for transformers
Pruning is an effective way to reduce the huge inference cost of Transformer models.
However, prior work on pruning Transformers requires retraining the models. This can add …
However, prior work on pruning Transformers requires retraining the models. This can add …
The state of sparsity in deep neural networks
We rigorously evaluate three state-of-the-art techniques for inducing sparsity in deep neural
networks on two large-scale learning tasks: Transformer trained on WMT 2014 English-to …
networks on two large-scale learning tasks: Transformer trained on WMT 2014 English-to …