Efficient acceleration of deep learning inference on resource-constrained edge devices: A review
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …
in breakthroughs in many areas. However, deploying these highly accurate models for data …
FPGA HLS today: successes, challenges, and opportunities
The year 2011 marked an important transition for FPGA high-level synthesis (HLS), as it
went from prototy** to deployment. A decade later, in this article, we assess the progress …
went from prototy** to deployment. A decade later, in this article, we assess the progress …
Dynamic neural networks: A survey
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …
models which have fixed computational graphs and parameters at the inference stage …
Squeezesegv3: Spatially-adaptive convolution for efficient point-cloud segmentation
LiDAR point-cloud segmentation is an important problem for many applications. For large-
scale point cloud segmentation, the de facto method is to project a 3D point cloud to get a …
scale point cloud segmentation, the de facto method is to project a 3D point cloud to get a …
Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search
Designing accurate and efficient ConvNets for mobile devices is challenging because the
design space is combinatorially large. Due to this, previous neural architecture search (NAS) …
design space is combinatorially large. Due to this, previous neural architecture search (NAS) …
Hawq: Hessian aware quantization of neural networks with mixed-precision
Abstract Model size and inference speed/power have become a major challenge in the
deployment of neural networks for many applications. A promising approach to address …
deployment of neural networks for many applications. A promising approach to address …
How do adam and training strategies help bnns optimization
Abstract The best performing Binary Neural Networks (BNNs) are usually attained using
Adam optimization and its multi-step training variants. However, to the best of our …
Adam optimization and its multi-step training variants. However, to the best of our …
Bit: Robustly binarized multi-distilled transformer
Modern pre-trained transformers have rapidly advanced the state-of-the-art in machine
learning, but have also grown in parameters and computational complexity, making them …
learning, but have also grown in parameters and computational complexity, making them …
Mix and match: A novel fpga-centric deep neural network quantization framework
Deep Neural Networks (DNNs) have achieved extraordinary performance in various
application domains. To support diverse DNN models, efficient implementations of DNN …
application domains. To support diverse DNN models, efficient implementations of DNN …
FracBNN: Accurate and FPGA-efficient binary neural networks with fractional activations
Binary neural networks (BNNs) have 1-bit weights and activations. Such networks are well
suited for FPGAs, as their dominant computations are bitwise arithmetic and the memory …
suited for FPGAs, as their dominant computations are bitwise arithmetic and the memory …