Challenges in high-performance computing
Abstract High-Performance Computing, HPC, has become one of the most active computer
science fields. Driven mainly by the need for high processing capabilities required by …
science fields. Driven mainly by the need for high processing capabilities required by …
Rcmp: Reconstructing RDMA-Based Memory Disaggregation via CXL
Memory disaggregation is a promising architecture for modern datacenters that separates
compute and memory resources into independent pools connected by ultra-fast networks …
compute and memory resources into independent pools connected by ultra-fast networks …
A quantitative approach for adopting disaggregated memory in HPC systems
Memory disaggregation has recently been adopted in data centers to improve resource
utilization, motivated by cost and sustainability. Recent studies on large-scale HPC facilities …
utilization, motivated by cost and sustainability. Recent studies on large-scale HPC facilities …
Exploring Numba and CuPy for GPU-Accelerated Monte Carlo Radiation Transport
This paper examines the performance of two popular GPU programming platforms, Numba
and CuPy, for Monte Carlo radiation transport calculations. We conducted tests involving …
and CuPy, for Monte Carlo radiation transport calculations. We conducted tests involving …
A survey of compute nodes with 100 TFLOPS and beyond for supercomputers
J Chang, K Lu, Y Guo, Y Wang, Z Zhao… - CCF Transactions on …, 2024 - Springer
With the Frontier supercomputer ranked first on the Top500 list, it marks the era of exascale
computing power for supercomputers, employing the compute nodes with double-precision …
computing power for supercomputers, employing the compute nodes with double-precision …
Shifting Between Compute and Memory Bounds: A Compression-Enabled Roofline Model
In the evolving landscape of high-performance computing, especially to fight the end of
Moore's Law and Dennard's Scaling, the ability to shift between compute-bound and …
Moore's Law and Dennard's Scaling, the ability to shift between compute-bound and …
ngAP: Non-blocking Large-scale Automata Processing on GPUs
Finite automata serve as compute kernels for various applications that require high
throughput. However, despite the increasing compute power of GPUs, their potential in …
throughput. However, despite the increasing compute power of GPUs, their potential in …
Exascale Quantum Mechanical Simulations: Navigating the Shifting Sands of Hardware and Software
The era of exascale computing presents both exciting opportunities and unique challenges
for quantum mechanical simulations. While the transition from petaflops to exascale …
for quantum mechanical simulations. While the transition from petaflops to exascale …
Data-Oriented Operating System for Big Data and Cloud.
Operating System (OS) is a critical piece of software that manages a computer's hardware
and resources, acting as the intermediary between the computer and the user. The existing …
and resources, acting as the intermediary between the computer and the user. The existing …
CPU-GPU Tuning for Modern Scientific Applications using Node-Level Heterogeneity
Scientific applications must be tuned for performance to run efficiently on supercomputers
having nodes with a CPU (or, a general-purpose host processor) and GPUs (or, accelerator …
having nodes with a CPU (or, a general-purpose host processor) and GPUs (or, accelerator …