A comprehensive survey on coded distributed computing: Fundamentals, challenges, and networking applications
Distributed computing has become a common approach for large-scale computation tasks
due to benefits such as high reliability, scalability, computation speed, and cost …
due to benefits such as high reliability, scalability, computation speed, and cost …
Short-dot: Computing large linear transforms distributedly using coded short dot products
Faced with saturation of Moore's law and increasing size and dimension of data, system
designers have increasingly resorted to parallel and distributed computing to reduce …
designers have increasingly resorted to parallel and distributed computing to reduce …
On the optimal recovery threshold of coded matrix multiplication
S Dutta, M Fahim, F Haddadpour… - IEEE Transactions …, 2019 - ieeexplore.ieee.org
We provide novel coded computation strategies for distributed matrix-matrix products that
outperform the recent “Polynomial code” constructions in recovery threshold, ie, the required …
outperform the recent “Polynomial code” constructions in recovery threshold, ie, the required …
Communication-computation efficient gradient coding
This paper develops coding techniques to reduce the running time of distributed learning
tasks. It characterizes the fundamental tradeoff to compute gradients in terms of three …
tasks. It characterizes the fundamental tradeoff to compute gradients in terms of three …
Entangled polynomial codes for secure, private, and batch distributed matrix multiplication: Breaking the" cubic" barrier
Q Yu, AS Avestimehr - 2020 IEEE International Symposium on …, 2020 - ieeexplore.ieee.org
In distributed matrix multiplication, a common scenario is to assign each worker a fraction of
the multiplication task, by partitioning the input matrices into smaller submatrices. In …
the multiplication task, by partitioning the input matrices into smaller submatrices. In …
A unified coded deep neural network training strategy based on generalized polydot codes
This paper has two main contributions. First, we propose a novel coding technique-
Generalized PolyDot-for matrix-vector products that advances on existing techniques for …
Generalized PolyDot-for matrix-vector products that advances on existing techniques for …
Cross subspace alignment codes for coded distributed batch computation
The goal of coded distributed computation is to efficiently distribute a computation task, such
as matrix multiplication, N-linear computation, or multivariate polynomial evaluation, across …
as matrix multiplication, N-linear computation, or multivariate polynomial evaluation, across …
A hierarchical incentive design toward motivating participation in coded federated learning
Federated Learning (FL) is a privacy-preserving collaborative learning approach that trains
artificial intelligence (AI) models without revealing local datasets of the FL workers. While FL …
artificial intelligence (AI) models without revealing local datasets of the FL workers. While FL …
Rateless codes for near-perfect load balancing in distributed matrix-vector multiplication
A Mallick, M Chaudhari, U Sheth… - Abstracts of the 2020 …, 2020 - dl.acm.org
Large-scale machine learning and data mining applications require computer systems to
perform massive matrix-vector and matrix-matrix multiplication operations that need to be …
perform massive matrix-vector and matrix-matrix multiplication operations that need to be …
Hierarchical coding for distributed computing
Coding for distributed computing supports low-latency computation by relieving the burden
of straggling workers. While most existing works assume a simple master-worker model, we …
of straggling workers. While most existing works assume a simple master-worker model, we …