scispace - formally typeset
Proceedings ArticleDOI

Presto: Edge-based Load Balancing for Fast Datacenter Networks

Reads0
Chats0
TLDR
A soft-edge load balancing scheme that closely tracks that of a single, non-blocking switch over many workloads and is adaptive to failures and topology asymmetry, called Presto is designed and implemented.
Abstract
Datacenter networks deal with a variety of workloads, ranging from latency-sensitive small flows to bandwidth-hungry large flows. Load balancing schemes based on flow hashing, e.g., ECMP, cause congestion when hash collisions occur and can perform poorly in asymmetric topologies. Recent proposals to load balance the network require centralized traffic engineering, multipath-aware transport, or expensive specialized hardware. We propose a mechanism that avoids these limitations by (i) pushing load-balancing functionality into the soft network edge (e.g., virtual switches) such that no changes are required in the transport layer, customer VMs, or networking hardware, and (ii) load balancing on fine-grained, near-uniform units of data (flowcells) that fit within end-host segment offload optimizations used to support fast networking speeds. We design and implement such a soft-edge load balancing scheme, called Presto, and evaluate it on a 10 Gbps physical testbed. We demonstrate the computational impact of packet reordering on receivers and propose a mechanism to handle reordering in the TCP receive offload functionality. Presto's performance closely tracks that of a single, non-blocking switch over many workloads and is adaptive to failures and topology asymmetry.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

HULA: Scalable Load Balancing Using Programmable Data Planes

TL;DR: HULA is presented, a data-plane load-balancing algorithm that outperforms a scalable extension to CONGA in average flow completion time and is designed for emerging programmable switches and programed in P4 to demonstrate that HULA could be run on such programmable chipsets, without requiring custom hardware.
Proceedings ArticleDOI

Re-architecting datacenter networks and stacks for low latency and high performance

TL;DR: NDP, a novel data-center transport architecture that achieves near-optimal completion times for short transfers and high flow throughput in a wide range of scenarios, including incast, is presented.
Proceedings ArticleDOI

Homa: a receiver-driven low-latency transport protocol using network priorities

TL;DR: Homa as discussed by the authors uses in-network priority queues to ensure low latency for short messages; priority allocation is managed dynamically by each receiver and integrated with a receiver-driven flow control mechanism.
Proceedings ArticleDOI

DRILL: Micro Load Balancing for Low-latency Data Center Networks

TL;DR: DRILL is presented, a datacenter fabric for Clos networks which performs micro load balancing to distribute load as evenly as possible on microsecond timescales and addresses the resulting key challenges of packet reordering and topological asymmetry.
Proceedings ArticleDOI

Resilient Datacenter Load Balancing in the Wild

TL;DR: Her Hermes is a datacenter load balancer that is resilient to the aforementioned uncertainties, and well handles uncertainties: under asymmetries, Hermes achieves up to 10% and 20% better flow completion time than CONGA and CLOVE; under switch failures, it outperforms all other schemes by over 32%.
References
More filters
Posted Content

A Quantitative Measure Of Fairness And Discrimination For Resource Allocation In Shared Computer Systems

TL;DR: A quantitative measure called Indiex of FRairness, applicable to any resource sharing or allocation problem, which is independent of the amount of the resource, and boundedness aids intuitive understanding of the fairness index.
Journal ArticleDOI

A scalable, commodity data center network architecture

TL;DR: This paper shows how to leverage largely commodity Ethernet switches to support the full aggregate bandwidth of clusters consisting of tens of thousands of elements and argues that appropriately architected and interconnected commodity switches may deliver more performance at less cost than available from today's higher-end solutions.
Proceedings ArticleDOI

VL2: a scalable and flexible data center network

TL;DR: VL2 is a practical network architecture that scales to support huge data centers with uniform high capacity between servers, performance isolation between services, and Ethernet layer-2 semantics, and is built on a working prototype.
Proceedings ArticleDOI

Network traffic characteristics of data centers in the wild

TL;DR: An empirical study of the network traffic in 10 data centers belonging to three different categories, including university, enterprise campus, and cloud data centers, which includes not only data centers employed by large online service providers offering Internet-facing applications but also data centers used to host data-intensive (MapReduce style) applications.
Journal ArticleDOI

CUBIC: a new TCP-friendly high-speed TCP variant

TL;DR: The CUBIC protocol modifies the linear window growth function of existing TCP standards to be a cubic function in order to improve the scalability of TCP over fast and long distance networks.
Related Papers (5)