Scheduling multi-tenant cloud workloads on accelerator-based systems
Citations
84 citations
Cites background from "Scheduling multi-tenant cloud workl..."
...Strings [Sengupta et al. 2014] extends its previous work, Rain [Sengupta et al. 2013], by including more effective scheduling polices....
[...]
...Strings [Sengupta et al. 2014] extends its previous work, Rain [Sengupta et al....
[...]
81 citations
Additional excerpts
...Categories and Subject Descriptors C.1.2 [Multiple Data Stream Architecture]: SingleInstruction, Multiple-Data Processors (SIMD); D.1.3 [ Programming Techniques]: Concurrent ProgrammingParallel Programming General Terms Design, Experimentation, Performance, Big Data Keywords Graph Analytics, Big Data, GPGPU, Performance Optimization, Data Movement Optimization ©2015 Association for Computing Machinery....
[...]
...…Descriptors C.1.2 [Multiple Data Stream Architecture]: SingleInstruction, Multiple-Data Processors (SIMD); D.1.3 [ Programming Techniques]: Concurrent ProgrammingParallel Programming General Terms Design, Experimentation, Performance, Big Data Keywords Graph Analytics, Big Data, GPGPU, Performance…...
[...]
72 citations
Cites background from "Scheduling multi-tenant cloud workl..."
...There is a large swath of related work on multi-tenancy for CPUs [36, 66, 97–103] and GPUs [29, 35, 37–42, 66, 104–110] due to its vitality for cloud-scale computing....
[...]
...In fact, the broader research community invested more than a decade of efforts to develop solutions across the computing stack to bring forth seamless and scalable multi-tenant cloud execution models [26–48]....
[...]
58 citations
Additional excerpts
...GraphReduce [22] framework can efficiently process graphs that cannot fit into the limited GPU memory [20,21] by mapping sub-graphs to the different memory abstractions of slow and fast memory [23]....
[...]
56 citations
Cites background from "Scheduling multi-tenant cloud workl..."
...presents the Strings scheduler which decouples CPU and GPU execution, and guides scheduling using feedback about execution time, GPU utilization, data transfer, and memory bandwidth utilization from each GPU [32]....
[...]
...Interference-aware scheduling on GPU servers has been attempted in the past [31, 32]....
[...]
...which enables a single GPU to be shared in both space and time [13, 32]....
[...]
References
4,476 citations
1,116 citations
1,064 citations
674 citations
619 citations