scispace - formally typeset
Journal ArticleDOI

Scheduling multithreaded computations by work stealing

TLDR
This paper gives the first provably good work-stealing scheduler for multithreaded computations with dependencies, and shows that the expected time to execute a fully strict computation on P processors using this scheduler is 1:1.
Abstract
This paper studies the problem of efficiently schedulling fully strict (i.e., well-structured) multithreaded computations on parallel computers. A popular and practical method of scheduling this kind of dynamic MIMD-style computation is “work stealing,” in which processors needing work steal computational threads from other processors. In this paper, we give the first provably good work-stealing scheduler for multithreaded computations with dependencies.Specifically, our analysis shows that the expected time to execute a fully strict computation on P processors using our work-stealing scheduler is T1/P + O(T ∞ , where T1 is the minimum serial execution time of the multithreaded computation and (T ∞ is the minimum execution time with an infinite number of processors. Moreover, the space required by the execution is at most S1P, where S1 is the minimum serial space requirement. We also show that the expected total communication of the algorithm is at most O(PT ∞( 1 + nd)Smax), where Smax is the size of the largest activation record of any thread and nd is the maximum number of times that any thread synchronizes with its parent. This communication bound justifies the folk wisdom that work-stealing schedulers are more communication efficient than their work-sharing counterparts. All three of these bounds are existentially optimal to within a constant factor.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Beyond nested parallelism: tight bounds on work-stealing overheads for parallel futures

TL;DR: It is shown that a broader metric, the number of deviations, is a better way to quantify work-stealing overhead for less restrictive forms of parallelism, including parallel futures, and identifies a family of applications that use futures and, in contrast to quicksort, incur significantly smaller overheads.
Journal ArticleDOI

Adaptive work-stealing with parallelism feedback

TL;DR: This work presents a randomized work-stealing thread scheduler for fork-join multithreaded jobs that provides continual parallelism feedback to the job scheduler in the form of requests for processors, and introduces a new technique called trim analysis, which allows it to prove that the adaptive thread Scheduler performs poorly on no more than a small number of time steps.
Proceedings ArticleDOI

Energy-efficient work-stealing language runtimes

TL;DR: The centerpiece of HERMES is two complementary algorithms to coordinate thread tempo: the workpath-sensitive algorithm determines tempo for each thread based on thief-victim relationships on the execution path, whereas the workload-sensitive algorithms selects appropriate tempo based on the size of work-stealing deques.
Journal ArticleDOI

On-the-Fly Pipeline Parallelism

TL;DR: A provably efficient scheduling algorithm, the Piper algorithm, is described, which integrates pipeline parallelism into a work-stealing scheduler, allowing pipeline and fork-join parallelism to be arbitrarily nested and automatically throttles the parallelism, precluding “runaway” pipelines.
Proceedings ArticleDOI

Deterministic parallel random-number generation for dynamic-multithreading platforms

TL;DR: This paper persuaded Intel to modify its commercial C/C++ compiler to include pedigrees, and built a library implementation of a deterministic parallel random-number generator called DotMix that compresses the pedigree and then "RC6-mixes" the result.
References
More filters
Journal ArticleDOI

Cilk: An Efficient Multithreaded Runtime System

TL;DR: It is shown that on real and synthetic applications, the “work” and “critical-path length” of a Cilk computation can be used to model performance accurately, and it is proved that for the class of “fully strict” (well-structured) programs, the Cilk scheduler achieves space, time, and communication bounds all within a constant factor of optimal.
Journal ArticleDOI

Bounds for certain multiprocessing anomalies

TL;DR: In this paper, precise bounds are derived for several anomalies of this type in a multiprocessing system composed of many identical processing units operating in parallel, and they show that an increase in the number of processing units can cause an increased total length of time needed to process a fixed set of tasks.
Proceedings ArticleDOI

The implementation of the Cilk-5 multithreaded language

TL;DR: Cilk-5's novel "two-clone" compilation strategy and its Dijkstra-like mutual-exclusion protocol for implementing the ready deque in the work-stealing scheduler are presented.
Journal ArticleDOI

The Parallel Evaluation of General Arithmetic Expressions

TL;DR: It is shown that arithmetic expressions with n ≥ 1 variables and constants; operations of addition, multiplication, and division; and any depth of parenthesis nesting can be evaluated in time 4 log 2 + 10(n - 1) using processors which can independently perform arithmetic operations in unit time.
Related Papers (5)