Deep learning based physical layer security for terrestrial communications in 5G and beyond networks: A survey
The key principle of physical layer security (PLS) is to permit the secure transmission of
confidential data using efficient signal-processing techniques. Also, deep learning (DL) has …
confidential data using efficient signal-processing techniques. Also, deep learning (DL) has …
An overview of the data-loader landscape: Comparative performance analysis
The efficiency of Deep Learning (DL) training jobs is critically dependent on dataloaders,
which facilitate the transfer of data from storage to DL-accelerated hardware during training …
which facilitate the transfer of data from storage to DL-accelerated hardware during training …
Analysis of deep learning libraries: Keras, pytorch, and MXnet
As many artificial neural libraries are develo** the deep learning algorithm and
implementing it became accessible to anyone. This study points out the disparity of …
implementing it became accessible to anyone. This study points out the disparity of …
A Survey on Spatio-temporal Big Data Analytics Ecosystem: Resource Management, Processing Platform, and Applications
With the rapid evolution of the Internet, Internet of Things (IoT), and geographic information
systems (GIS), spatio-temporal Big Data (STBD) is experiencing exponential growth …
systems (GIS), spatio-temporal Big Data (STBD) is experiencing exponential growth …
xCCL: A Survey of Industry-Led Collective Communication Libraries for Deep Learning
Abstract Machine learning techniques have become ubiquitous both in industry and
academic applications. Increasing model sizes and training data volumes necessitate fast …
academic applications. Increasing model sizes and training data volumes necessitate fast …
Accelerating CPU-based distributed DNN training on modern HPC clusters using bluefield-2 DPUs
The Deep Learning (DL) training process consists of multiple phases—data augmentation,
training, and validation of the trained model. Traditionally, these phases are executed either …
training, and validation of the trained model. Traditionally, these phases are executed either …
The mit supercloud dataset
Artificial intelligence (AI) and Machine learning (ML) workloads are an increasingly larger
share of the compute workloads in traditional High-Performance Computing (HPC) centers …
share of the compute workloads in traditional High-Performance Computing (HPC) centers …
An optimized error-controlled mpi collective framework integrated with lossy compression
With the ever-increasing computing power of supercomputers and the growing scale of
scientific applications, the efficiency of MPI collective communications turns out to be a …
scientific applications, the efficiency of MPI collective communications turns out to be a …
Parallel and distributed training of deep neural networks: A brief overview
Deep neural networks and deep learning are becoming important and popular techniques in
modern services and applications. The training of these networks is computationally …
modern services and applications. The training of these networks is computationally …
gzccl: Compression-accelerated collective communication framework for gpu clusters
GPU-aware collective communication has become a major bottleneck for modern computing
platforms as GPU computing power rapidly rises. A traditional approach is to directly …
platforms as GPU computing power rapidly rises. A traditional approach is to directly …