Towards asynchronous federated learning based threat detection: A DC-Adam approach

P Tian, Z Chen, W Yu, W Liao - Computers & Security, 2021 - Elsevier
The increasing popularity and widespread use of Internet of Things (IoT) and Cyber-Physical
Systems (CPS) technologies have produced a significant need for the integration of cloud …

Web traffic time series forecasting using LSTM neural networks with distributed asynchronous training

R Casado-Vara, A Martin del Rey, D Pérez-Palau… - Mathematics, 2021 - mdpi.com
Evaluating web traffic on a web server is highly critical for web service providers since,
without a proper demand forecast, customers could have lengthy waiting times and abandon …

Federated reinforcement learning for training control policies on multiple IoT devices

HK Lim, JB Kim, JS Heo, YH Han - Sensors, 2020 - mdpi.com
Reinforcement learning has recently been studied in various fields and also used to
optimally control IoT devices supporting the expansion of Internet connection beyond the …

Profiling dnn workloads on a volta-based dgx-1 system

SA Mojumder, MS Louis, Y Sun… - 2018 IEEE …, 2018 - ieeexplore.ieee.org
High performance multi-GPU systems are widely used to accelerate training of deep neural
networks (DNNs) by exploiting the inherently massive parallel nature of the training process …

Speeding up deep learning with transient servers

S Li, RJ Walls, L Xu, T Guo - 2019 IEEE International …, 2019 - ieeexplore.ieee.org
Distributed training frameworks, like TensorFlow, have been proposed as a means to reduce
the training time of deep learning models by using a cluster of GPU servers. While such …

Staleness analysis in asynchronous optimization

H Al-Lawati, S Draper - IEEE Transactions on Signal and …, 2022 - ieeexplore.ieee.org
Distributed optimization is widely used to solve large-scale optimization problems by
parallelizing gradient-based algorithms across multiple computing nodes. In asynchronous …

Making asynchronous stochastic gradient descent work for transformers

AF Aji, K Heafield - arxiv preprint arxiv:1906.03496, 2019 - arxiv.org
Asynchronous stochastic gradient descent (SGD) is attractive from a speed perspective
because workers do not wait for synchronization. However, the Transformer model …

Gradient staleness in asynchronous optimization under random communication delays

H Al-Lawati, SC Draper - ICASSP 2022-2022 IEEE …, 2022 - ieeexplore.ieee.org
Distributed optimization is widely used to solve large-scale optimization problems by
parallelizing gradient-based algorithms across multiple computing nodes. In asynchronous …

[PDF][PDF] Approximating neural machine translation for efficiency

AF Aji - 2020 - era.ed.ac.uk
Neural machine translation (NMT) has been shown to outperform statistical machine
translation. However, NMT models typically require a large number of parameters and are …

Accelerating the processing of deep neural networks

J Li - 2020 - wrap.warwick.ac.uk
Artificial Intelligent (AI) has become the most potent and forward-looking force in the
technologies over the past decade. It allowed breakthrough applications that have truly …