Scaling for edge inference of deep neural networks

X Xu, Y Ding, SX Hu, M Niemier, J Cong, Y Hu… - Nature Electronics, 2018 - nature.com
Deep neural networks offer considerable potential across a range of applications, from
advanced manufacturing to autonomous cars. A clear trend in deep neural networks is the …

Testability and dependability of AI hardware: Survey, trends, challenges, and perspectives

F Su, C Liu, HG Stratigopoulos - IEEE Design & Test, 2023 - ieeexplore.ieee.org
Hardware realization of artificial intelligence (AI) requires new design styles and even
underlying technologies than those used in traditional digital processors or logic circuits …

CONV-SRAM: An energy-efficient SRAM with in-memory dot-product computation for low-power convolutional neural networks

A Biswas, AP Chandrakasan - IEEE Journal of Solid-State …, 2018 - ieeexplore.ieee.org
This paper presents an energy-efficient static random access memory (SRAM) with
embedded dot-product computation capability, for binary-weight convolutional neural …

UNPU: An energy-efficient deep neural network accelerator with fully variable weight bit precision

J Lee, C Kim, S Kang, D Shin, S Kim… - IEEE Journal of Solid …, 2018 - ieeexplore.ieee.org
An energy-efficient deep neural network (DNN) accelerator, unified neural processing unit
(UNPU), is proposed for mobile deep learning applications. The UNPU can support both …

C3SRAM: An in-memory-computing SRAM macro based on robust capacitive coupling computing mechanism

Z Jiang, S Yin, JS Seo, M Seok - IEEE Journal of Solid-State …, 2020 - ieeexplore.ieee.org
This article presents C3SRAM, an in-memory-computing SRAM macro. The macro is an
SRAM module with the circuits embedded in bitcells and peripherals to perform hardware …

A 0.086-mm 12.7-pJ/SOP 64k-Synapse 256-Neuron Online-Learning Digital Spiking Neuromorphic Processor in 28-nm CMOS

C Frenkel, M Lefebvre, JD Legat… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
Shifting computing architectures from von Neumann to event-based spiking neural networks
(SNNs) uncovers new opportunities for low-power processing of sensory data in …

Ares: A framework for quantifying the resilience of deep neural networks

B Reagen, U Gupta, L Pentecost… - Proceedings of the 55th …, 2018 - dl.acm.org
As the use of deep neural networks continues to grow, so does the fraction of compute
cycles devoted to their execution. This has led the CAD and architecture communities to …

Spiking neural networks hardware implementations and challenges: A survey

M Bouvier, A Valentian, T Mesquida… - ACM Journal on …, 2019 - dl.acm.org
Neuromorphic computing is henceforth a major research field for both academic and
industrial actors. As opposed to Von Neumann machines, brain-inspired processors aim at …

CirCNN: accelerating and compressing deep neural networks using block-circulant weight matrices

C Ding, S Liao, Y Wang, Z Li, N Liu, Y Zhuo… - Proceedings of the 50th …, 2017 - dl.acm.org
Large-scale deep neural networks (DNNs) are both compute and memory intensive. As the
size of DNNs continues to grow, it is critical to improve the energy efficiency and …

Tensordimm: A practical near-memory processing architecture for embeddings and tensor operations in deep learning

Y Kwon, Y Lee, M Rhu - Proceedings of the 52nd Annual IEEE/ACM …, 2019 - dl.acm.org
Recent studies from several hyperscalars pinpoint to embedding layers as the most memory-
intensive deep learning (DL) algorithm being deployed in today's datacenters. This paper …