Fusion of memristor and digital compute-in-memory processing for energy-efficient edge computing

TH Wen, JM Hung, WH Huang, CJ Jhang, YC Lo… - Science, 2024 - science.org
Artificial intelligence (AI) edge devices prefer employing high-capacity nonvolatile compute-
in-memory (CIM) to achieve high energy efficiency and rapid wakeup-to-response with …

A nonvolatile Al-edge processor with 4MB SLC-MLC hybrid-mode ReRAM compute-in-memory macro and 51.4-251TOPS/W

WH Huang, TH Wen, JM Hung… - … Solid-State Circuits …, 2023 - ieeexplore.ieee.org
Low-power Al edge devices should provide short-latency (T_WK-RP) and low-energy
(E_WK-RP) wakeup responses from power-off mode to handle event-triggered computing …

TinyVers: A tiny versatile system-on-chip with state-retentive eMRAM for ML inference at the extreme edge

V Jain, S Giraldo, J De Roose, L Mei… - IEEE Journal of Solid …, 2023 - ieeexplore.ieee.org
Extreme edge devices or Internet-of-Things (IoT) nodes require both ultra-low power (ULP)
always-on (AON) processing as well as the ability to do on-demand sampling and …

Defines: Enabling fast exploration of the depth-first scheduling space for dnn accelerators through analytical modeling

L Mei, K Goetschalckx, A Symons… - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
DNN workloads can be scheduled onto DNN accelerators in many different ways: from layer-
by-layer scheduling to cross-layer depth-first scheduling (aka layer fusion, or cascaded …

A nonvolatile AI-edge processor with SLC–MLC hybrid ReRAM compute-in-memory macro using current–voltage-hybrid readout scheme

HH Hsu, TH Wen, WH Huang, WS Khwa… - IEEE Journal of Solid …, 2023 - ieeexplore.ieee.org
On-chip non-volatile compute-in-memory (nvCIM) enables artificial intelligence (AI)-edge
processors to perform multiply-and-accumulate (MAC) operations while enabling the non …

A 22 nm Floating-Point ReRAM Compute-in-Memory Macro Using Residue-Shared ADC for AI Edge Device

HH Hsu, TH Wen, WS Khwa, WH Huang… - IEEE Journal of Solid …, 2024 - ieeexplore.ieee.org
Artificial intelligence (AI) edge devices increasingly require the enhanced accuracy of
floating-point (FP) multiply-and-accumulate (MAC) operations as well as nonvolatile on-chip …

ML processors are going multi-core: A performance dream or a scheduling nightmare?

M Verhelst, M Shi, L Mei - IEEE Solid-State Circuits Magazine, 2022 - ieeexplore.ieee.org
Applications of machine learning (ML) increasingly penetrate into our daily routines, our
work, and our living environments. In this way, more complex machine intelligence …

HTVM: Efficient neural network deployment on heterogeneous TinyML platforms

J Van Delm, M Vandersteegen… - 2023 60th ACM/IEEE …, 2023 - ieeexplore.ieee.org
Optimal deployment of deep neural networks (DNNs) on state-of-the-art Systems-on-Chips
(SoCs) is crucial for tiny machine learning (TinyML) at the edge. The complexity of these …

PATRONoC: Parallel AXI transport reducing overhead for networks-on-chip targeting multi-accelerator DNN platforms at the edge

V Jain, M Cavalcante, N Bruschi… - 2023 60th ACM/IEEE …, 2023 - ieeexplore.ieee.org
Emerging deep neural network (DNN) applications require high-performance multi-core
hardware acceleration with large data bursts. Classical network-on-chips (NoCs) use serial …

Pianissimo: A Sub-mW Class DNN Accelerator With Progressively Adjustable Bit-Precision

J Suzuki, J Yu, M Yasunaga, ÁL García-Arias… - IEEE …, 2023 - ieeexplore.ieee.org
With the widespread adoption of edge AI, the diversity of application requirements and
fluctuating computational demands present significant challenges. Conventional …