Proceedings ArticleDOI
Presto: Edge-based Load Balancing for Fast Datacenter Networks
Keqiang He,Eric J. Rozner,Kanak B. Agarwal,Wesley M. Felter,John B. Carter,Aditya Akella +5 more
- Vol. 45, Iss: 4, pp 465-478
Reads0
Chats0
TLDR
A soft-edge load balancing scheme that closely tracks that of a single, non-blocking switch over many workloads and is adaptive to failures and topology asymmetry, called Presto is designed and implemented.Abstract:
Datacenter networks deal with a variety of workloads, ranging from latency-sensitive small flows to bandwidth-hungry large flows. Load balancing schemes based on flow hashing, e.g., ECMP, cause congestion when hash collisions occur and can perform poorly in asymmetric topologies. Recent proposals to load balance the network require centralized traffic engineering, multipath-aware transport, or expensive specialized hardware. We propose a mechanism that avoids these limitations by (i) pushing load-balancing functionality into the soft network edge (e.g., virtual switches) such that no changes are required in the transport layer, customer VMs, or networking hardware, and (ii) load balancing on fine-grained, near-uniform units of data (flowcells) that fit within end-host segment offload optimizations used to support fast networking speeds. We design and implement such a soft-edge load balancing scheme, called Presto, and evaluate it on a 10 Gbps physical testbed. We demonstrate the computational impact of packet reordering on receivers and propose a mechanism to handle reordering in the TCP receive offload functionality. Presto's performance closely tracks that of a single, non-blocking switch over many workloads and is adaptive to failures and topology asymmetry.read more
Citations
More filters
Proceedings ArticleDOI
HULA: Scalable Load Balancing Using Programmable Data Planes
TL;DR: HULA is presented, a data-plane load-balancing algorithm that outperforms a scalable extension to CONGA in average flow completion time and is designed for emerging programmable switches and programed in P4 to demonstrate that HULA could be run on such programmable chipsets, without requiring custom hardware.
Proceedings ArticleDOI
Re-architecting datacenter networks and stacks for low latency and high performance
Mark Handley,Costin Raiciu,Alexandru Agache,Andrei Voinescu,Andrew W. Moore,Gianni Antichi,Marcin Wójcik +6 more
TL;DR: NDP, a novel data-center transport architecture that achieves near-optimal completion times for short transfers and high flow throughput in a wide range of scenarios, including incast, is presented.
Proceedings ArticleDOI
Homa: a receiver-driven low-latency transport protocol using network priorities
TL;DR: Homa as discussed by the authors uses in-network priority queues to ensure low latency for short messages; priority allocation is managed dynamically by each receiver and integrated with a receiver-driven flow control mechanism.
Proceedings ArticleDOI
DRILL: Micro Load Balancing for Low-latency Data Center Networks
TL;DR: DRILL is presented, a datacenter fabric for Clos networks which performs micro load balancing to distribute load as evenly as possible on microsecond timescales and addresses the resulting key challenges of packet reordering and topological asymmetry.
Proceedings ArticleDOI
Resilient Datacenter Load Balancing in the Wild
TL;DR: Her Hermes is a datacenter load balancer that is resilient to the aforementioned uncertainties, and well handles uncertainties: under asymmetries, Hermes achieves up to 10% and 20% better flow completion time than CONGA and CLOVE; under switch failures, it outperforms all other schemes by over 32%.
References
More filters
Posted Content
A Quantitative Measure Of Fairness And Discrimination For Resource Allocation In Shared Computer Systems
Raj Jain,Dah Ming Chiu,W. Hawe +2 more
TL;DR: A quantitative measure called Indiex of FRairness, applicable to any resource sharing or allocation problem, which is independent of the amount of the resource, and boundedness aids intuitive understanding of the fairness index.
Journal ArticleDOI
A scalable, commodity data center network architecture
TL;DR: This paper shows how to leverage largely commodity Ethernet switches to support the full aggregate bandwidth of clusters consisting of tens of thousands of elements and argues that appropriately architected and interconnected commodity switches may deliver more performance at less cost than available from today's higher-end solutions.
Proceedings ArticleDOI
VL2: a scalable and flexible data center network
Albert Greenberg,James R. Hamilton,Navendu Jain,Srikanth Kandula,Changhoon Kim,Parantap Lahiri,David A. Maltz,Parveen Patel,Sudipta Sengupta +8 more
TL;DR: VL2 is a practical network architecture that scales to support huge data centers with uniform high capacity between servers, performance isolation between services, and Ethernet layer-2 semantics, and is built on a working prototype.
Proceedings ArticleDOI
Network traffic characteristics of data centers in the wild
TL;DR: An empirical study of the network traffic in 10 data centers belonging to three different categories, including university, enterprise campus, and cloud data centers, which includes not only data centers employed by large online service providers offering Internet-facing applications but also data centers used to host data-intensive (MapReduce style) applications.
Journal ArticleDOI
CUBIC: a new TCP-friendly high-speed TCP variant
Sangtae Ha,Injong Rhee,Lisong Xu +2 more
TL;DR: The CUBIC protocol modifies the linear window growth function of existing TCP standards to be a cubic function in order to improve the scalability of TCP over fast and long distance networks.