scispace - formally typeset
Open AccessProceedings Article

Polynomial codes: an optimal design for high-dimensional coded matrix multiplication

Reads0
Chats0
TLDR
This work considers a large-scale matrix multiplication problem where the computation is carried out using a distributed system with a master node and multiple worker nodes, where each worker can store parts of the input matrices, and proposes a computation strategy that leverages ideas from coding theory to design intermediate computations at the worker nodes to efficiently deal with straggling workers.
Abstract
We consider a large-scale matrix multiplication problem where the computation is carried out using a distributed system with a master node and multiple worker nodes, where each worker can store parts of the input matrices. We propose a computation strategy that leverages ideas from coding theory to design intermediate computations at the worker nodes, in order to optimally deal with straggling workers. The proposed strategy, named as polynomial codes, achieves the optimum recovery threshold, defined as the minimum number of workers that the master needs to wait for in order to compute the output. This is the first code that achieves the optimal utilization of redundancy for tolerating stragglers or failures in distributed matrix multiplication. Furthermore, by leveraging the algebraic structure of polynomial codes, we can map the reconstruction problem of the final output to a polynomial interpolation problem, which can be solved efficiently. Polynomial codes provide order-wise improvement over the state of the art in terms of recovery threshold, and are also optimal in terms of several other metrics including computation latency and communication load. Moreover, we extend this code to distributed convolution and show its order-wise optimality.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

A Fundamental Tradeoff Between Computation and Communication in Distributed Computing

TL;DR: A coded scheme, named “coded distributed computing” (CDC), is proposed to demonstrate that increasing the computation load of the Map functions by a factor of r can create novel coding opportunities that reduce the communication load by the same factor.
Posted Content

A Fundamental Tradeoff between Computation and Communication in Distributed Computing

TL;DR: In this article, a coded distributed computing (CDC) scheme is proposed to reduce the communication load in distributed computing, where the overall computation is decomposed into computing a set of Map and Reduce functions distributedly across multiple computing nodes.
Journal ArticleDOI

On the Optimal Recovery Threshold of Coded Matrix Multiplication

TL;DR: Novel coded computation strategies for distributed matrix–matrix products that outperform the recent “Polynomial code” constructions in recovery threshold, i.e., the required number of successful workers are provided.
Proceedings ArticleDOI

Improving Distributed Gradient Descent Using Reed-Solomon Codes

TL;DR: In this article, the authors adopt the framework of Tandon et al. and present a deterministic scheme that, for a prescribed per-machine computational effort, recovers the gradient from the least number of machines $f$ theoretically permissible, via an O(f 2 ) decoding algorithm.
Journal ArticleDOI

Straggler Mitigation in Distributed Matrix Multiplication: Fundamental Limits and Optimal Coding

TL;DR: While evaluating bilinear complexity is a well-known challenging problem, it is shown that optimal recovery threshold for linear coding strategies can be approximated within a factor of 2 of this fundamental quantity.
References
More filters
Journal ArticleDOI

MapReduce: simplified data processing on large clusters

TL;DR: This paper presents the implementation of MapReduce, a programming model and an associated implementation for processing and generating large data sets that runs on a large cluster of commodity machines and is highly scalable.
Proceedings Article

Spark: cluster computing with working sets

TL;DR: Spark can outperform Hadoop by 10x in iterative machine learning jobs, and can be used to interactively query a 39 GB dataset with sub-second response time.
Proceedings ArticleDOI

Improving MapReduce performance in heterogeneous environments

TL;DR: A new scheduling algorithm, Longest Approximate Time to End (LATE), that is highly robust to heterogeneity and can improve Hadoop response times by a factor of 2 in clusters of 200 virtual machines on EC2.
Journal ArticleDOI

The tail at scale

TL;DR: Software techniques that tolerate latency variability are vital to building responsive large-scale Web services.
Journal ArticleDOI

Algorithm-Based Fault Tolerance for Matrix Operations

TL;DR: Algorithm-based fault tolerance schemes are proposed to detect and correct errors when matrix operations such as addition, multiplication, scalar product, LU-decomposition, and transposition are performed using multiple processor systems.
Related Papers (5)