Design possibilities and challenges of DNN models: a review on the perspective of end devices
Abstract Deep Neural Network (DNN) models for both resource-rich environments and
resource-constrained devices have become abundant in recent years. As of now, the …
resource-constrained devices have become abundant in recent years. As of now, the …
Matraptor: A sparse-sparse matrix multiplication accelerator based on row-wise product
Sparse-sparse matrix multiplication (SpGEMM) is a computation kernel widely used in
numerous application domains such as data analytics, graph processing, and scientific …
numerous application domains such as data analytics, graph processing, and scientific …
Sanger: A co-design framework for enabling sparse attention using reconfigurable architecture
In recent years, attention-based models have achieved impressive performance in natural
language processing and computer vision applications by effectively capturing contextual …
language processing and computer vision applications by effectively capturing contextual …
Approximate computing: Concepts, architectures, challenges, applications, and future directions
AM Dalloo, AJ Humaidi, AK Al Mhdawi… - IEEE …, 2024 - ieeexplore.ieee.org
The unprecedented progress in computational technologies led to a substantial proliferation
of artificial intelligence applications, notably in the era of big data and IoT devices. In the …
of artificial intelligence applications, notably in the era of big data and IoT devices. In the …
Tensaurus: A versatile accelerator for mixed sparse-dense tensor computations
N Srivastava, H **, S Smith, H Rong… - … Symposium on High …, 2020 - ieeexplore.ieee.org
Tensor factorizations are powerful tools in many machine learning and data analytics
applications. Tensors are often sparse, which makes sparse tensor factorizations memory …
applications. Tensors are often sparse, which makes sparse tensor factorizations memory …
Hardware acceleration of sparse and irregular tensor computations of ml models: A survey and insights
Machine learning (ML) models are widely used in many important domains. For efficiently
processing these computational-and memory-intensive applications, tensors of these …
processing these computational-and memory-intensive applications, tensors of these …
An efficient hardware accelerator for structured sparse convolutional neural networks on FPGAs
Deep convolutional neural networks (CNNs) have achieved state-of-the-art performance in a
wide range of applications. However, deeper CNN models, which are usually computation …
wide range of applications. However, deeper CNN models, which are usually computation …
Hardware implementation of 1D-CNN architecture for ECG arrhythmia classification
Electrocardiography (ECG) has been used as a diagnostic tool for various heart diseases. It
is most effective in detecting myocardial infarction and fatal arrhythmias. This work proposes …
is most effective in detecting myocardial infarction and fatal arrhythmias. This work proposes …
The lottery ticket hypothesis for object recognition
Recognition tasks, such as object recognition and keypoint estimation, have seen
widespread adoption in recent years. Most state-of-the-art methods for these tasks use deep …
widespread adoption in recent years. Most state-of-the-art methods for these tasks use deep …
Highlight: Efficient and flexible dnn acceleration with hierarchical structured sparsity
Due to complex interactions among various deep neural network (DNN) optimization
techniques, modern DNNs can have weights and activations that are dense or sparse with …
techniques, modern DNNs can have weights and activations that are dense or sparse with …