Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
BitWave: Exploiting column-based bit-level sparsity for deep learning acceleration
Bit-serial computation facilitates bit-wise sequential data processing, offering numerous
benefits, such as a reduced area footprint and dynamically-adaptive computational …
benefits, such as a reduced area footprint and dynamically-adaptive computational …
Optimus: An operator fusion framework for deep neural networks
The reduction of neural parameters and operations for the applications on embedded and
IoT platforms in current deep neural network (DNN) architectures has received increasing …
IoT platforms in current deep neural network (DNN) architectures has received increasing …
Cascading structured pruning: enabling high data reuse for sparse dnn accelerators
Performance and efficiency of running modern Deep Neural Networks (DNNs) are heavily
bounded by data movement. To mitigate the data movement bottlenecks, recent DNN …
bounded by data movement. To mitigate the data movement bottlenecks, recent DNN …
[ΒΙΒΛΙΟ][B] Low-power computer vision: improve the efficiency of artificial intelligence
Energy efficiency is critical for running computer vision on battery-powered systems, such as
mobile phones or UAVs (unmanned aerial vehicles, or drones). This book collects the …
mobile phones or UAVs (unmanned aerial vehicles, or drones). This book collects the …
DSLR-CNN: Efficient CNN Acceleration using Digit-Serial Left-to-Right Arithmetic
Digit-serial arithmetic has emerged as a viable approach for designing hardware
accelerators, reducing interconnections, area utilization, and power consumption. However …
accelerators, reducing interconnections, area utilization, and power consumption. However …
Optimus: towards optimal layer-fusion on deep learning processors
Neural network layer fusion has been proposed to parallelize the inference of neural layers
and thus significantly reduces the feature-induced memory accesses. However, how to fuse …
and thus significantly reduces the feature-induced memory accesses. However, how to fuse …
ASBP: Automatic structured bit-pruning for RRAM-based NN accelerator
Network sparsity or pruning is an extensively studied method to optimize the computation
efficiency of deep neural networks (DNNs) for CMOS-based accelerators, such as FPGAs …
efficiency of deep neural networks (DNNs) for CMOS-based accelerators, such as FPGAs …
Msd: Mixing signed digit representations for hardware-efficient dnn acceleration on fpga with heterogeneous resources
By quantizing weights with different precision for different parts of a network, mixed-precision
quantization promises to reduce the hardware cost and improve the speed of deep neural …
quantization promises to reduce the hardware cost and improve the speed of deep neural …
Special session: Fault-tolerant deep learning: A hierarchical perspective
With the rapid advancements of deep learning in the past decade, it can be foreseen that
deep learning will be continuously deployed in more and more safety-critical applications …
deep learning will be continuously deployed in more and more safety-critical applications …
Bit-balance: Model-hardware codesign for accelerating nns by exploiting bit-level sparsity
W Sun, Z Zou, D Liu, W Sun, S Chen… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Bit-serial architectures can handle Neural Networks (NNs) with different weight precision,
achieving higher resource efficiency compared with bit-parallel architectures. Besides, the …
achieving higher resource efficiency compared with bit-parallel architectures. Besides, the …