Homa: a receiver-driven low-latency transport protocol using network priorities

@article{Montazeri2018HomaAR,
  title={Homa: a receiver-driven low-latency transport protocol using network priorities},
  author={Behnam Montazeri and Yilong Li and Mohammad Alizadeh and John K. Ousterhout},
  journal={Proceedings of the 2018 Conference of the ACM Special Interest Group on Data Communication},
  year={2018}
}
Homa is a new transport protocol for datacenter networks. It provides exceptionally low latency, especially for workloads with a high volume of very short messages, and it also supports large messages and high network utilization. Homa uses in-network priority queues to ensure low latency for short messages; priority allocation is managed dynamically by each receiver and integrated with a receiver-driven flow control mechanism. Homa also uses controlled overcommitment of receiver downlinks to… Expand
Polo: Receiver-Driven Congestion Control for Low Latency over Commodity Network Fabric
TLDR
Polo is presented to realize low latency for flows over commodity network fabric relying on Explicit Congestion Notification (ECN) and priority queues and results show that Polo outperforms the state-of-art receiver-driven protocols in a wide range of scenarios including incast. Expand
Expanding across time to deliver bandwidth efficiency and low latency
TLDR
Opera is presented, a dynamic network that delivers latency-sensitive traffic quickly by relying on multi-hop forwarding in the same way as expander-graph-based approaches, but provides near-optimal bandwidth for bulk flows through direct forwarding over time-varying source-to-destination circuits. Expand
AMRT: Anti-ECN Marking to Improve Utilization of Receiver-driven Transmission in Data Center
TLDR
A new receiver-driven transport scheme using anti-ECN (Explicit Congestion Notification) marking to achieve both near-zero queueing delay and full link utilization by reasonably increasing sending rate in the case of under-utilization. Expand
Exploring Token-Oriented In-Network Prioritization in Datacenter Networks
TLDR
This article proposes a readily-deployable remedy to achieve in-network prioritization by pushing both switch and end-host hardware capacity to an extreme end and implements a running TOP system with Linux hosts and commodity switches, and investigates the applicability of TOP. Expand
Packet Order Matters! Improving Application Performance by Deliberately Delaying Packets
Data centers increasingly deploy commodity servers with high-speed network interfaces to enable low-latency communication. However, achieving low latency at high data rates crucially depends on howExpand
Chameleon: predictable latency and high utilization with queue-aware and adaptive source routing
TLDR
Compared to state-of-the-art, this paper finds that Chameleon can admit and embed significantly, i.e., up to 15 times more flows, improving network utilization while meeting strict latency guarantees. Expand
PL2: Towards Predictable Low Latency in Rack-Scale Networks
TLDR
A Predictable Low Latency (PL2) network architecture for rack-scale systems with Ethernet as interconnecting fabric that leverages programmable Ethernet switches to carefully schedule packets such that they incur no loss with NIC and switch queues maintained at small, near-zero levels. Expand
FatPaths: Routing in Supercomputers, Data Centers, and Clouds with Low-Diameter Networks when Shortest Paths Fall Short
We introduce FatPaths: a simple, generic, and robust routing architecture for Ethernet stacks. FatPaths enables state-of-the-art low-diameter topologies such as Slim Fly to achieve unprecedentedExpand
Enabling Programmable Transport Protocols in High-Speed NICs
TLDR
This paper enables programmable transport protocols in high-speed NICs by designing Tonic, a flexible hardware architecture for transport logic that satisfies constraints while being programmable with a simple API. Expand
MDTCP: Towards a Practical Multipath Transport Protocol for Telco Cloud Datacenters
TLDR
This work proposes a multipath transport for telco cloud datacenters called coupled multipath datacenter TCP, MDTCP, which evolves MPTCP subflows to employ ECN signals to react to congestion before queue overflow, offering both reduced latency and higher network utilization. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 36 REFERENCES
Re-architecting datacenter networks and stacks for low latency and high performance
TLDR
NDP, a novel data-center transport architecture that achieves near-optimal completion times for short transfers and high flow throughput in a wide range of scenarios, including incast, is presented. Expand
Less Is More: Trading a Little Bandwidth for Ultra-Low Latency in the Data Center
TLDR
The HULL (High-bandwidth Ultra-Low Latency) architecture is presented to balance two seemingly contradictory goals: near baseline fabric latency and high bandwidth utilization and results show that by sacrificing a small amount of bandwidth, HULL can dramatically reduce average and tail latencies in the data center. Expand
Fastpass: A Centralized “Zero-Queue” Datacenter Network
TLDR
This paper describes Fastpass, a datacenter network architecture built using this principle that achieves high throughput comparable to current networks at a 240x reduction is queue lengths, and achieves much fairer and consistent flow throughputs than the baseline TCP. Expand
Congestion Control for Large-Scale RDMA Deployments
TLDR
DCQCN, an end-to-end congestion control scheme for RoCEv2, is introduced and it is shown that DCQCN dramatically improves throughput and fairness of Ro CEv2 RDMA traffic. Expand
Queues Don't Matter When You Can JUMP Them!
TLDR
It is shown that QJUMP achieves bounded latency and reduces in-network interference by up to 300×, outperforming Ethernet Flow Control (802.3x), ECN (WRED) and DCTCP and pFabric. Expand
Better never than late: meeting deadlines in datacenter networks
The soft real-time nature of large scale web applications in today's datacenters, combined with their distributed workflow, leads to deadlines being associated with the datacenter applicationExpand
Presto: Edge-based Load Balancing for Fast Datacenter Networks
TLDR
A soft-edge load balancing scheme that closely tracks that of a single, non-blocking switch over many workloads and is adaptive to failures and topology asymmetry, called Presto is designed and implemented. Expand
Data center TCP (DCTCP)
TLDR
DCTCP enables the applications to handle 10X the current background traffic, without impacting foreground traffic, thus largely eliminating incast problems, and delivers the same or better throughput than TCP, while using 90% less buffer space. Expand
Finishing flows quickly with preemptive scheduling
TLDR
It is demonstrated that PDQ significantly outperforms TCP, RCP and D3 in data center environments, and is stable, resilient to packet loss, and preserves nearly all its performance gains even given inaccurate flow information. Expand
Credit-Scheduled Delay-Bounded Congestion Control for Datacenters
TLDR
This paper presents a new radical approach, called ExpressPass, an end-to-end credit-scheduled, delay-bounded congestion control for datacenters that greatly improves performance under heavy incast workloads and significantly reduces the flow completion times. Expand
...
1
2
3
4
...