Intriguing properties of quantization at scale

A Ahmadian, S Dash, H Chen… - Advances in …, 2023 - proceedings.neurips.cc
Emergent properties have been widely adopted as a term to describe behavior not present
in smaller models but observed in larger models (Wei et al., 2022a). Recent work suggests …

Codg-reram: An algorithm-hardware co-design to accelerate semi-structured gnns on reram

Y Luo, P Behnam, K Thorat, Z Liu… - 2022 IEEE 40th …, 2022 - ieeexplore.ieee.org
Graph Neural Networks (GCNs) have attracted wide attention and are applied to the real
world. However, due to the ever-growing graph data with significant irregularities, off-chip …

Inshrinkerator: Compressing Deep Learning Training Checkpoints via Dynamic Quantization

A Agrawal, S Reddy, S Bhattamishra… - Proceedings of the …, 2024 - dl.acm.org
The likelihood of encountering in-training failures rises substantially with larger Deep
Learning (DL) training workloads, leading to lost work and resource wastage. Such failures …

Harmonica: Hybrid Accelerator to Overcome Imperfections of Mixed-signal DNN Accelerators

P Behnam, U Kamal, A Shafiee… - 2024 IEEE …, 2024 - ieeexplore.ieee.org
In recent years, PIM-based mixed-signal accelerators have been proposed as energy-and
area-efficient solutions with ultra-high throughput to accelerate DNN computations …

Dynaquant: Compressing deep learning training checkpoints via dynamic quantization

A Agrawal, S Reddy, S Bhattamishra… - arxiv e …, 2023 - ui.adsabs.harvard.edu
With the increase in the scale of Deep Learning (DL) training workloads in terms of compute
resources and time consumption, the likelihood of encountering in-training failures rises …

BWA-NIMC: Budget-based Workload Allocation for Hybrid Near/In-Memory-Computing

CT Huang, CY Chang, YC Chuang… - 2023 60th ACM/IEEE …, 2023 - ieeexplore.ieee.org
To enable efficient computation for convolutional neural networks, in-memory-computing
(IMC) is proposed to perform computation within memory. However, the non-ideality …

An algorithm-hardware co-design framework to overcome imperfections of mixed-signal dnn accelerators

P Behnam, U Kamal, S Mukhopadhyay - arxiv preprint arxiv:2208.13896, 2022 - arxiv.org
In recent years, processing in memory (PIM) based mixedsignal designs have been
proposed as energy-and area-efficient solutions with ultra high throughput to accelerate …

Reliable edge intelligence in unreliable environment

M Lee, X She, B Chakraborty, S Dash… - … , Automation & Test …, 2021 - ieeexplore.ieee.org
A key challenge for deployment of artificial intelligence (AI) in real-time safety-critical
systems at the edge is to ensure reliable performance even in unreliable environments. This …

DEA-NIMC: Dynamic Energy-Aware Policy for Near/In-Memory Computing Hybrid Architecture

YC Wu, CT Huang, AYA Wu - 2023 IEEE 36th International …, 2023 - ieeexplore.ieee.org
In-memory computing (IMC) has become the current trend to accelerate the inference of
deep neural networks (DNNs). Nonetheless, IMC suffers from variations that significantly …

Hessian-Based Parameter Quantization Method for BERT

W Byun, S Mukhopadhyay - 2023 IEEE 66th International …, 2023 - ieeexplore.ieee.org
Transformer-based language models have shown outstanding performance in various NLP
tasks, but using them on edge devices is very challenging due to their notorious memory …