Dataperf: Benchmarks for data-centric ai development
Abstract Machine learning research has long focused on models rather than datasets, and
prominent datasets are used for common ML tasks without regard to the breadth, difficulty …
prominent datasets are used for common ML tasks without regard to the breadth, difficulty …
A survey of on-device machine learning: An algorithms and learning theory perspective
The predominant paradigm for using machine learning models on a device is to train a
model in the cloud and perform inference using the trained model on the device. However …
model in the cloud and perform inference using the trained model on the device. However …
Priority-based parameter propagation for distributed DNN training
Data parallel training is widely used for scaling distributed deep neural network (DNN)
training. However, the performance benefits are often limited by the communication-heavy …
training. However, the performance benefits are often limited by the communication-heavy …
Gist: Efficient data encoding for deep neural network training
Modern deep neural networks (DNNs) training typically relies on GPUs to train complex
hundred-layer deep networks. A significant problem facing both researchers and industry …
hundred-layer deep networks. A significant problem facing both researchers and industry …
Analysis of dawnbench, a time-to-accuracy machine learning performance benchmark
Researchers have proposed hardware, software, and algorithmic optimizations to improve
the computational performance of deep learning. While some of these optimizations perform …
the computational performance of deep learning. While some of these optimizations perform …
Parameter hub: a rack-scale parameter server for distributed deep neural network training
Distributed deep neural network (DDNN) training constitutes an increasingly important
workload that frequently runs in the cloud. Larger DNN models and faster compute engines …
workload that frequently runs in the cloud. Larger DNN models and faster compute engines …
An overview of the data-loader landscape: Comparative performance analysis
The efficiency of Deep Learning (DL) training jobs is critically dependent on dataloaders,
which facilitate the transfer of data from storage to DL-accelerated hardware during training …
which facilitate the transfer of data from storage to DL-accelerated hardware during training …
DLBench: a comprehensive experimental evaluation of deep learning frameworks
Deep Learning (DL) has achieved remarkable progress over the last decade on various
tasks such as image recognition, speech recognition, and natural language processing. In …
tasks such as image recognition, speech recognition, and natural language processing. In …
A modular benchmarking infrastructure for high-performance and reproducible deep learning
We introduce Deep500: the first customizable benchmarking infrastructure that enables fair
comparison of the plethora of deep learning frameworks, algorithms, libraries, and …
comparison of the plethora of deep learning frameworks, algorithms, libraries, and …
[HTML][HTML] Effect of neural network structure in accelerating performance and accuracy of a convolutional neural network with GPU/TPU for image analytics
Background In deep learning the most significant breakthrough in the field of image
recognition, object detection language processing was done by Convolutional Neural …
recognition, object detection language processing was done by Convolutional Neural …