Scalable deep learning on distributed infrastructures: Challenges, techniques, and tools

R Mayer, HA Jacobsen - ACM Computing Surveys (CSUR), 2020 - dl.acm.org
Deep Learning (DL) has had an immense success in the recent past, leading to state-of-the-
art results in various domains, such as image recognition and natural language processing …

Towards secure intrusion detection systems using deep learning techniques: Comprehensive analysis and review

SW Lee, M Mohammadi, S Rashidi… - Journal of Network and …, 2021 - Elsevier
Providing a high-performance Intrusion Detection System (IDS) can be very effective in
controlling malicious behaviors and cyber-attacks. Regarding the ever-growing negative …

Deep leakage from gradients

L Zhu, Z Liu, S Han - Advances in neural information …, 2019 - proceedings.neurips.cc
Passing gradient is a widely used scheme in modern multi-node learning system (eg,
distributed training, collaborative learning). In a long time, people used to believe that …

Neural tangents: Fast and easy infinite neural networks in python

R Novak, L **ao, J Hron, J Lee, AA Alemi… - arxiv preprint arxiv …, 2019 - arxiv.org
Neural Tangents is a library designed to enable research into infinite-width neural networks.
It provides a high-level API for specifying complex and hierarchical neural network …

Exascale deep learning for climate analytics

T Kurth, S Treichler, J Romero… - … conference for high …, 2018 - ieeexplore.ieee.org
We extract pixel-level masks of extreme weather patterns using variants of Tiramisu and
DeepLabv3+ neural networks. We describe improvements to the software frameworks, input …

Extremely large minibatch sgd: Training resnet-50 on imagenet in 15 minutes

T Akiba, S Suzuki, K Fukuda - arxiv preprint arxiv:1711.04325, 2017 - arxiv.org
We demonstrate that training ResNet-50 on ImageNet for 90 epochs can be achieved in 15
minutes with 1024 Tesla P100 GPUs. This was made possible by using a large minibatch …

Bigdl: A distributed deep learning framework for big data

JJ Dai, Y Wang, X Qiu, D Ding, Y Zhang… - Proceedings of the …, 2019 - dl.acm.org
ThispaperpresentsBigDL (adistributeddeeplearning framework for Apache Spark), which
has been used by a variety of users in the industry for building deep learning applications on …

Train sparsely, generate densely: Memory-efficient unsupervised training of high-resolution temporal gan

M Saito, S Saito, M Koyama, S Kobayashi - International Journal of …, 2020 - Springer
Training of generative adversarial network (GAN) on a video dataset is a challenge because
of the sheer size of the dataset and the complexity of each observation. In general, the …

The case for in-network computing on demand

Y Tokusashi, HT Dang, F Pedone, R Soulé… - Proceedings of the …, 2019 - dl.acm.org
Programmable network hardware can run services traditionally deployed on servers,
resulting in orders-of-magnitude improvements in performance. Yet, despite these …

Large-scale distributed second-order optimization using kronecker-factored approximate curvature for deep convolutional neural networks

K Osawa, Y Tsuji, Y Ueno, A Naruse… - Proceedings of the …, 2019 - openaccess.thecvf.com
Large-scale distributed training of deep neural networks suffers from the generalization gap
caused by the increase in the effective mini-batch size. Previous approaches try to solve this …