Proceedings of the 2018 Afternoon Workshop on Kernel Bypassing Networks 2018
DOI: 10.1145/3229538.3229541
|View full text |Cite
|
Sign up to set email alerts
|

Flow control for Latency-Critical RPCs

Abstract: In today's modern datacenters, the waiting time spent within a server's queue is a major contributor of the end-to-end tail latency of µs-scale remote procedure calls. In traditional TCP, congestion control handles in-network congestion, while flow control was designed to avoid memory overruns in streaming scenarios. The latter is unfortunately oblivious to the load on the server when processing short requests from multiple clients at very high rates. Acknowledging flow control as the mechanism that controls q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 39 publications
0
2
0
Order By: Relevance
“…The load generated by devices tends to follow certain rules (e.g., Poisson inter-arrival time distribution [37]). At the beginning of each time slice, the end device possesses multiple tasks k m (t) and the data size of k m (t) obeys the uniform distribution.…”
Section: Mobile Device Nodementioning
confidence: 99%
“…The load generated by devices tends to follow certain rules (e.g., Poisson inter-arrival time distribution [37]). At the beginning of each time slice, the end device possesses multiple tasks k m (t) and the data size of k m (t) obeys the uniform distribution.…”
Section: Mobile Device Nodementioning
confidence: 99%
“…Provisioning a receive buffer of limited size on the server requires the transport protocol to signal a "failure to deliver" (NACK) if the request is dropped because of a full queue. It is up to the client to react to a NACK reception; for example, the request could be retried or sent to a different server, as proposed by Kogias et al [45]. Exposing delivery failures to the client follows the end-to-end principle in systems design [46]: the client application is best equipped to handle such violations and should be informed immediately.…”
Section: Bounded Server-side Queuingmentioning
confidence: 99%
“…The need for low latency has led systems designers to aggressively limit queuing in the transport and RPC protocol stacks themselves [45], [59]. Kogias et al [45] also observe that limiting server-side queuing is critical for µs-scale RPCs, and use TCP flow control to limit the number of requests per connection based on the application's SLO. NEBULA performs buffer management for hardware-rather than software-terminated protocols.…”
Section: Related Workmentioning
confidence: 99%