A modern primer on processing in memory
Modern computing systems are overwhelmingly designed to move data to computation. This
design choice goes directly against at least three key trends in computing that cause …
design choice goes directly against at least three key trends in computing that cause …
Benchmarking a new paradigm: Experimental analysis and characterization of a real processing-in-memory system
Many modern workloads, such as neural networks, databases, and graph processing, are
fundamentally memory-bound. For such workloads, the data movement between main …
fundamentally memory-bound. For such workloads, the data movement between main …
Breaking the von Neumann bottleneck: architecture-level processing-in-memory technology
The “memory wall” problem or so-called von Neumann bottleneck limits the efficiency of
conventional computer architectures, which move data from memory to CPU for …
conventional computer architectures, which move data from memory to CPU for …
Rowhammer: A retrospective
This retrospective paper describes the RowHammer problem in dynamic random access
memory (DRAM), which was initially introduced by Kim et al. at the ISCA 2014 Conference …
memory (DRAM), which was initially introduced by Kim et al. at the ISCA 2014 Conference …
Processing-in-memory: A workload-driven perspective
Many modern and emerging applications must process increasingly large volumes of data.
Unfortunately, prevalent computing paradigms are not designed to efficiently handle such …
Unfortunately, prevalent computing paradigms are not designed to efficiently handle such …
DAMOV: A new methodology and benchmark suite for evaluating data movement bottlenecks
Data movement between the CPU and main memory is a first-order obstacle against improv
ing performance, scalability, and energy efficiency in modern systems. Computer systems …
ing performance, scalability, and energy efficiency in modern systems. Computer systems …
FPGA-based near-memory acceleration of modern data-intensive applications
Modern data-intensive applications demand high computational capabilities with strict
power constraints. Unfortunately, such applications suffer from a significant waste of both …
power constraints. Unfortunately, such applications suffer from a significant waste of both …
Benchmarking a new paradigm: An experimental analysis of a real processing-in-memory architecture
Many modern workloads, such as neural networks, databases, and graph processing, are
fundamentally memory-bound. For such workloads, the data movement between main …
fundamentally memory-bound. For such workloads, the data movement between main …
Syncron: Efficient synchronization support for near-data-processing architectures
Near-Data-Processing (NDP) architectures present a promising way to alleviate data
movement costs and can provide significant performance and energy benefits to parallel …
movement costs and can provide significant performance and energy benefits to parallel …
NERO: A near high-bandwidth memory stencil accelerator for weather prediction modeling
Ongoing climate change calls for fast and accurate weather and climate modeling. However,
when solving large-scale weather prediction simulations, state-of-the-art CPU and GPU …
when solving large-scale weather prediction simulations, state-of-the-art CPU and GPU …