Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Mitigating Write Disturbance in Non-Volatile Memory via Coupling Machine Learning with Out-of-Place Updates
Non-volatile memory (NVM) opens up new opportunities to resolve scaling restrictions of
main memory, yet it is still hindered by the write disturbance (WD) problem. The WD problem …
main memory, yet it is still hindered by the write disturbance (WD) problem. The WD problem …
VeriGOOD-ML: An open-source flow for automated ML hardware synthesis
H Esmaeilzadeh, S Ghodrati, J Gu… - 2021 IEEE/ACM …, 2021 - ieeexplore.ieee.org
This paper introduces VeriGOOD-ML, an automated methodology for generating Verilog
with no human in the loop, starting from a high-level description of a machine learning (ML) …
with no human in the loop, starting from a high-level description of a machine learning (ML) …
A reschedulable dataflow-SIMD execution for increased utilization in CGRA cross-domain acceleration
C Yin, N **g, J Jiang, Q Wang… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
When a coarse-grained reconfigurable array (CGRA) architecture shifts toward cross-
domain acceleration, control flow and memory accesses often degrade the processing …
domain acceleration, control flow and memory accesses often degrade the processing …
Energy-efficient hardware acceleration of shallow machine learning applications
Z Zeng, SS Sapatnekar - 2023 Design, Automation & Test in …, 2023 - ieeexplore.ieee.org
ML accelerators have largely focused on building general platforms for deep neural
networks (DNNs), but less so on shallow machine learning (SML) algorithms. This paper …
networks (DNNs), but less so on shallow machine learning (SML) algorithms. This paper …
LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale
Recently, there has been an extensive research effort in building efficient large language
model (LLM) inference serving systems. These efforts not only include innovations in the …
model (LLM) inference serving systems. These efforts not only include innovations in the …
End-to-end synthesis of dynamically controlled machine learning accelerators
Edge systems are required to autonomously make real-time decisions based on large
quantities of input data under strict power, performance, area, and other constraints. Meeting …
quantities of input data under strict power, performance, area, and other constraints. Meeting …
Improving utilization of dataflow unit for multi-batch processing
Dataflow architectures can achieve much better performance and higher efficiency than
general-purpose core, approaching the performance of a specialized design while retaining …
general-purpose core, approaching the performance of a specialized design while retaining …
A 28-nm Software-Defined Accelerator Chip With Circuit-Pipeline Scaling and Intrinsic Physical Unclonable Function Enabling Secure Configuration
J Zhu, B Yang, L Chen, J Chen, Y Zhang… - IEEE Journal of Solid …, 2025 - ieeexplore.ieee.org
As emerging applications raise ever-boosting and varying computational demand, the
reconfigurable accelerator is becoming prevalent due to balanced performance, efficiency …
reconfigurable accelerator is becoming prevalent due to balanced performance, efficiency …
Workload-Aware Hardware Accelerator Mining for Distributed Deep Learning Training
In this paper, we present a novel technique to search for hardware architectures of
accelerators optimized for end-to-end training of deep neural networks (DNNs). Our …
accelerators optimized for end-to-end training of deep neural networks (DNNs). Our …
APPEND: Rethinking ASIP Synthesis in the Era of AI
Application-specific instruction-set processors (ASIP) has been widely used to speedup
specific applications based on general-purpose processor (CPU) ISA-extension and …
specific applications based on general-purpose processor (CPU) ISA-extension and …