A review of convolutional neural network architectures and their optimizations

S Cong, Y Zhou - Artificial Intelligence Review, 2023 - Springer
The research advances concerning the typical architectures of convolutional neural
networks (CNNs) as well as their optimizations are analyzed and elaborated in detail in this …

A crossbar array of magnetoresistive memory devices for in-memory computing

S Jung, H Lee, S Myung, H Kim, SK Yoon, SW Kwon… - Nature, 2022 - nature.com
Implementations of artificial neural networks that borrow analogue techniques could
potentially offer low-power alternatives to fully digital approaches,–. One notable example is …

A systematic literature review on binary neural networks

R Sayed, H Azmi, H Shawkey, AH Khalil… - IEEE Access, 2023 - ieeexplore.ieee.org
This paper presents an extensive literature review on Binary Neural Network (BNN). BNN
utilizes binary weights and activation function parameters to substitute the full-precision …

A survey of quantization methods for efficient neural network inference

A Gholami, S Kim, Z Dong, Z Yao… - Low-power computer …, 2022 - taylorfrancis.com
This chapter provides approaches to the problem of quantizing the numerical values in deep
Neural Network computations, covering the advantages/disadvantages of current methods …

Pruning and quantization for deep neural network acceleration: A survey

T Liang, J Glossner, L Wang, S Shi, X Zhang - Neurocomputing, 2021 - Elsevier
Deep neural networks have been applied in many applications exhibiting extraordinary
abilities in the field of computer vision. However, complex network architectures challenge …

Up or down? adaptive rounding for post-training quantization

M Nagel, RA Amjad, M Van Baalen… - International …, 2020 - proceedings.mlr.press
When quantizing neural networks, assigning each floating-point weight to its nearest fixed-
point value is the predominant approach. We find that, perhaps surprisingly, this is not the …

Binary neural networks: A survey

H Qin, R Gong, X Liu, X Bai, J Song, N Sebe - Pattern Recognition, 2020 - Elsevier
The binary neural network, largely saving the storage and computation, serves as a
promising technique for deploying deep models on resource-limited devices. However, the …

Differentiable soft quantization: Bridging full-precision and low-bit neural networks

R Gong, X Liu, S Jiang, T Li, P Hu… - Proceedings of the …, 2019 - openaccess.thecvf.com
Hardware-friendly network quantization (eg, binary/uniform quantization) can efficiently
accelerate the inference and meanwhile reduce memory consumption of the deep neural …

Learning to quantize deep networks by optimizing quantization intervals with task loss

S Jung, C Son, S Lee, J Son, JJ Han… - Proceedings of the …, 2019 - openaccess.thecvf.com
Reducing bit-widths of activations and weights of deep networks makes it efficient to
compute and store them in memory, which is crucial in their deployments to resource-limited …

Network quantization with element-wise gradient scaling

J Lee, D Kim, B Ham - … of the IEEE/CVF conference on …, 2021 - openaccess.thecvf.com
Network quantization aims at reducing bit-widths of weights and/or activations, particularly
important for implementing deep neural networks with limited hardware resources. Most …