Coded computation is an emerging research area that leverages concepts from erasure coding to mitigate the effect of stragglers (slow nodes) in distributed computation clusters, especially for matrix computation problems. In this work, we present a class of distributed matrix-vector multiplication schemes that are based on codes in the Rosenbloom-Tsfasman metric and universally decodable matrices. Our schemes take into account the inherent computation order within a worker node. In particular, they allow us to effectively leverage partial computations performed by stragglers (a feature that many prior works lack). An additional main contribution of our work is a companion matrix-based embedding of these codes that allows us to obtain sparse...
Coded computation techniques provide robustness against straggling workers in distributed computing....
A ubiquitous problem in computer science research is the optimization of computation on large data s...
Large matrix multiplications commonly take place in large-scale machine-learning applications. Often...
Distributed computing systems are well-known to suffer from the problem of slow or failed nodes; the...
Distributed matrix computations (matrix-vector and matrix-matrix multiplications) are at the heart o...
Distributed matrix multiplication is widely used in several scientific domains. It is well recognize...
The current BigData era routinely requires the processing of large scale data on massive distributed...
In distributed computing systems, it is well recognized that worker nodes that are slow (called stra...
The overall execution time of distributed matrix computations is often dominated by slow worker node...
Coded computing is an effective technique to mitigate “stragglers” in large-scale and distributed ma...
Matrix multiplication is a fundamental building block in many machine learning models. As the input ...
We propose two coded schemes for the distributed computing problem of multiplying a matrix by a set ...
Existing approaches to distributed matrix computations involve allocating coded combinations of subm...
Polynomial coding has been proposed as a solution to the straggler mitigation problem in distributed...
Polynomial coding has been proposed as a solution to the straggler mitigation problem in distributed...
Coded computation techniques provide robustness against straggling workers in distributed computing....
A ubiquitous problem in computer science research is the optimization of computation on large data s...
Large matrix multiplications commonly take place in large-scale machine-learning applications. Often...
Distributed computing systems are well-known to suffer from the problem of slow or failed nodes; the...
Distributed matrix computations (matrix-vector and matrix-matrix multiplications) are at the heart o...
Distributed matrix multiplication is widely used in several scientific domains. It is well recognize...
The current BigData era routinely requires the processing of large scale data on massive distributed...
In distributed computing systems, it is well recognized that worker nodes that are slow (called stra...
The overall execution time of distributed matrix computations is often dominated by slow worker node...
Coded computing is an effective technique to mitigate “stragglers” in large-scale and distributed ma...
Matrix multiplication is a fundamental building block in many machine learning models. As the input ...
We propose two coded schemes for the distributed computing problem of multiplying a matrix by a set ...
Existing approaches to distributed matrix computations involve allocating coded combinations of subm...
Polynomial coding has been proposed as a solution to the straggler mitigation problem in distributed...
Polynomial coding has been proposed as a solution to the straggler mitigation problem in distributed...
Coded computation techniques provide robustness against straggling workers in distributed computing....
A ubiquitous problem in computer science research is the optimization of computation on large data s...
Large matrix multiplications commonly take place in large-scale machine-learning applications. Often...