A comprehensive survey on coded distributed computing: Fundamentals, challenges, and networking applications
Distributed computing has become a common approach for large-scale computation tasks
due to benefits such as high reliability, scalability, computation speed, and cost …
due to benefits such as high reliability, scalability, computation speed, and cost …
Cross subspace alignment codes for coded distributed batch computation
The goal of coded distributed computation is to efficiently distribute a computation task, such
as matrix multiplication, N-linear computation, or multivariate polynomial evaluation, across …
as matrix multiplication, N-linear computation, or multivariate polynomial evaluation, across …
Coding for large-scale distributed machine learning
This article aims to give a comprehensive and rigorous review of the principles and recent
development of coding for large-scale distributed machine learning (DML). With increasing …
development of coding for large-scale distributed machine learning (DML). With increasing …
Wireless MapReduce distributed computing
Motivated by mobile edge computing and wireless data centers, we study a wireless
distributed computing framework where the distributed nodes exchange information over a …
distributed computing framework where the distributed nodes exchange information over a …
GCSA codes with noise alignment for secure coded multi-party batch matrix multiplication
A secure multi-party batch matrix multiplication problem (SMBMM) is considered, where the
goal is to allow a master to efficiently compute the pairwise products of two batches of …
goal is to allow a master to efficiently compute the pairwise products of two batches of …
Coded distributed computing with partial recovery
Coded computation techniques provide robustness against straggling workers in distributed
computing. However, most of the existing schemes require exact provisioning of the …
computing. However, most of the existing schemes require exact provisioning of the …
Straggler-aware distributed learning: Communication–computation latency trade-off
When gradient descent (GD) is scaled to many parallel workers for large-scale machine
learning applications, its per-iteration computation time is limited by straggling workers …
learning applications, its per-iteration computation time is limited by straggling workers …
Erasurehead: Distributed gradient descent without delays using approximate gradient coding
We present ErasureHead, a new approach for distributed gradient descent (GD) that
mitigates system delays by employing approximate gradient coding. Gradient coded …
mitigates system delays by employing approximate gradient coding. Gradient coded …
Distributed gradient descent with coded partial gradient computations
Coded computation techniques provide robustness against straggling servers in distributed
computing, with the following limitations: First, they increase decoding complexity. Second …
computing, with the following limitations: First, they increase decoding complexity. Second …
Storage-computation-communication tradeoff in distributed computing: Fundamental limits and complexity
Distributed computing has become one of the most important frameworks in dealing with
large computation tasks. In this paper, we propose a systematic construction of coded …
large computation tasks. In this paper, we propose a systematic construction of coded …