A systematic literature review on binary neural networks

R Sayed, H Azmi, H Shawkey, AH Khalil… - IEEE Access, 2023 - ieeexplore.ieee.org
This paper presents an extensive literature review on Binary Neural Network (BNN). BNN
utilizes binary weights and activation function parameters to substitute the full-precision …

A survey of quantization methods for efficient neural network inference

A Gholami, S Kim, Z Dong, Z Yao… - Low-Power Computer …, 2022 - taylorfrancis.com
This chapter provides approaches to the problem of quantizing the numerical values in deep
Neural Network computations, covering the advantages/disadvantages of current methods …

Binary neural networks: A survey

H Qin, R Gong, X Liu, X Bai, J Song, N Sebe - Pattern Recognition, 2020 - Elsevier
The binary neural network, largely saving the storage and computation, serves as a
promising technique for deploying deep models on resource-limited devices. However, the …

Forward and backward information retention for accurate binary neural networks

H Qin, R Gong, X Liu, M Shen, Z Wei… - Proceedings of the …, 2020 - openaccess.thecvf.com
Weight and activation binarization is an effective approach to deep neural network
compression and can accelerate the inference by leveraging bitwise operations. Although …

Training binary neural networks with real-to-binary convolutions

B Martinez, J Yang, A Bulat, G Tzimiropoulos - arxiv preprint arxiv …, 2020 - arxiv.org
This paper shows how to train binary networks to within a few percent points ($\sim 3-5\% $)
of the full precision counterpart. We first show how to build a strong baseline, which already …

How do adam and training strategies help bnns optimization

Z Liu, Z Shen, S Li, K Helwegen… - International …, 2021 - proceedings.mlr.press
Abstract The best performing Binary Neural Networks (BNNs) are usually attained using
Adam optimization and its multi-step training variants. However, to the best of our …

Recu: Reviving the dead weights in binary neural networks

Z Xu, M Lin, J Liu, J Chen, L Shao… - Proceedings of the …, 2021 - openaccess.thecvf.com
Binary neural networks (BNNs) have received increasing attention due to their superior
reductions of computation and memory. Most existing works focus on either lessening the …

FracBNN: Accurate and FPGA-efficient binary neural networks with fractional activations

Y Zhang, J Pan, X Liu, H Chen, D Chen… - The 2021 ACM/SIGDA …, 2021 - dl.acm.org
Binary neural networks (BNNs) have 1-bit weights and activations. Such networks are well
suited for FPGAs, as their dominant computations are bitwise arithmetic and the memory …

Latent weights do not exist: Rethinking binarized neural network optimization

K Helwegen, J Widdicombe, L Geiger… - Advances in neural …, 2019 - proceedings.neurips.cc
Abstract Optimization of Binarized Neural Networks (BNNs) currently relies on real-valued
latent weights to accumulate small update steps. In this paper, we argue that these latent …

Pokebnn: A binary pursuit of lightweight accuracy

Y Zhang, Z Zhang, L Lew - … of the IEEE/CVF Conference on …, 2022 - openaccess.thecvf.com
Abstract Optimization of Top-1 ImageNet promotes enormous networks that may be
impractical in inference settings. Binary neural networks (BNNs) have the potential to …