Mlperf training benchmark

P Mattson, C Cheng, G Diamos… - Proceedings of …, 2020 - proceedings.mlsys.org
Abstract Machine learning is experiencing an explosion of software and hardware solutions,
and needs industry-standard performance benchmarks to drive design and enable …

Smartsage: training large-scale graph neural networks using in-storage processing architectures

Y Lee, J Chung, M Rhu - Proceedings of the 49th Annual International …, 2022 - dl.acm.org
Graph neural networks (GNNs) can extract features by learning both the representation of
each objects (ie, graph nodes) and the relationship across different objects (ie, the edges …

McDRAM v2: In-dynamic random access memory systolic array accelerator to address the large model problem in deep neural networks on the edge

S Cho, H Choi, E Park, H Shin, S Yoo - IEEE Access, 2020 - ieeexplore.ieee.org
The energy efficiency of accelerating hundreds of MB-large deep neural networks (DNNs) in
a mobile environment is less than that of a server-class big chip accelerator because of the …

Develo** a recommendation benchmark for mlperf training and inference

CJ Wu, R Burke, EH Chi, J Konstan, J McAuley… - arxiv preprint arxiv …, 2020 - arxiv.org
Deep learning-based recommendation models are used pervasively and broadly, for
example, to recommend movies, products, or other information most relevant to users, in …

Understanding the implication of non-volatile memory for large-scale graph neural network training

Y Lee, Y Kwon, M Rhu - IEEE Computer Architecture Letters, 2021 - ieeexplore.ieee.org
Graph neural networks (GNNs) can extract features by learning both the representation of
each objects (ie, graph nodes) as well as the relationship across different objects (ie, the …