Pqa: Exploring the potential of product quantization in dnn hardware acceleration

A Abouelhamayed, A Cui… - ACM Transactions on …, 2024 - dl.acm.org
Conventional multiply-accumulate (MAC) operations have long dominated computation time
for deep neural networks (DNNs), especially convolutional neural networks (CNNs) …

Power-aware training for energy-efficient printed neuromorphic circuits

H Zhao, P Pal, M Hefenbrock, M Beigl… - 2023 IEEE/ACM …, 2023 - ieeexplore.ieee.org
There is an increasing demand for next-generation flexible electronics in emerging low-cost
applications such as smart packaging and smart bandages, where conventional silicon …

Stella Nera: Achieving 161 TOp/s/W with Multiplier-free DNN Acceleration based on Approximate Matrix Multiplication

J Schönleber, L Cavigelli, R Andri, M Perotti… - arxiv preprint arxiv …, 2023 - arxiv.org
From classical HPC to deep learning, MatMul is at the heart of today's computing. The recent
Maddness method approximates MatMul without the need for multiplication by using a hash …

Multiplication-Free Lookup-Based CNN Accelerator using Residual Vector Quantization and Its FPGA Implementation

H Fuketa, T Katashita, Y Hori, M Hioki - IEEE Access, 2024 - ieeexplore.ieee.org
In this paper, a table lookup-based computing technique is proposed to perform
convolutional neural network (CNN) inference without multiplication, and its FPGA …

Full-stack optimization for cam-only dnn inference

JPC de Lima, AA Khan, L Carro… - … Design, Automation & …, 2024 - ieeexplore.ieee.org
The accuracy of neural networks has greatly improved across various domains over the past
years. Their ever-increasing complexity, however, leads to prohibitively high energy …

LUT-DLA: Lookup Table as Efficient Extreme Low-Bit Deep Learning Accelerator

G Li, S Ye, C Chen, Y Wang, F Yang, T Cao… - arxiv preprint arxiv …, 2025 - arxiv.org
The emergence of neural network capabilities invariably leads to a significant surge in
computational demands due to expanding model sizes and increased computational …

LUTIN: Efficient Neural Network Inference with Table Lookup

SZ Lin, YC Chen, YH Chang, TW Kuo… - Proceedings of the 29th …, 2024 - dl.acm.org
DNN models are becoming increasingly large and complex, but they are also being
deployed on commodity devices that require low power and latency but lack specialized …