Efficient acceleration of deep learning inference on resource-constrained edge devices: A review

MMH Shuvo, SK Islam, J Cheng… - Proceedings of the …, 2022 - ieeexplore.ieee.org
Successful integration of deep neural networks (DNNs) or deep learning (DL) has resulted
in breakthroughs in many areas. However, deploying these highly accurate models for data …

FPGA HLS today: successes, challenges, and opportunities

J Cong, J Lau, G Liu, S Neuendorffer, P Pan… - ACM Transactions on …, 2022 - dl.acm.org
The year 2011 marked an important transition for FPGA high-level synthesis (HLS), as it
went from prototy** to deployment. A decade later, in this article, we assess the progress …

Dynamic neural networks: A survey

Y Han, G Huang, S Song, L Yang… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Dynamic neural network is an emerging research topic in deep learning. Compared to static
models which have fixed computational graphs and parameters at the inference stage …

Squeezesegv3: Spatially-adaptive convolution for efficient point-cloud segmentation

C Xu, B Wu, Z Wang, W Zhan, P Vajda… - Computer Vision–ECCV …, 2020 - Springer
LiDAR point-cloud segmentation is an important problem for many applications. For large-
scale point cloud segmentation, the de facto method is to project a 3D point cloud to get a …

Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search

B Wu, X Dai, P Zhang, Y Wang, F Sun… - Proceedings of the …, 2019 - openaccess.thecvf.com
Designing accurate and efficient ConvNets for mobile devices is challenging because the
design space is combinatorially large. Due to this, previous neural architecture search (NAS) …

Hawq: Hessian aware quantization of neural networks with mixed-precision

Z Dong, Z Yao, A Gholami… - Proceedings of the …, 2019 - openaccess.thecvf.com
Abstract Model size and inference speed/power have become a major challenge in the
deployment of neural networks for many applications. A promising approach to address …

How do adam and training strategies help bnns optimization

Z Liu, Z Shen, S Li, K Helwegen… - International …, 2021 - proceedings.mlr.press
Abstract The best performing Binary Neural Networks (BNNs) are usually attained using
Adam optimization and its multi-step training variants. However, to the best of our …

Bit: Robustly binarized multi-distilled transformer

Z Liu, B Oguz, A Pappu, L **ao, S Yih… - Advances in neural …, 2022 - proceedings.neurips.cc
Modern pre-trained transformers have rapidly advanced the state-of-the-art in machine
learning, but have also grown in parameters and computational complexity, making them …

Mix and match: A novel fpga-centric deep neural network quantization framework

SE Chang, Y Li, M Sun, R Shi, HKH So… - … Symposium on High …, 2021 - ieeexplore.ieee.org
Deep Neural Networks (DNNs) have achieved extraordinary performance in various
application domains. To support diverse DNN models, efficient implementations of DNN …

FracBNN: Accurate and FPGA-efficient binary neural networks with fractional activations

Y Zhang, J Pan, X Liu, H Chen, D Chen… - The 2021 ACM/SIGDA …, 2021 - dl.acm.org
Binary neural networks (BNNs) have 1-bit weights and activations. Such networks are well
suited for FPGAs, as their dominant computations are bitwise arithmetic and the memory …