Datacenter traffic control: Understanding techniques and tradeoffs

M Noormohammadpour… - … Surveys & Tutorials, 2017 - ieeexplore.ieee.org
Datacenters provide cost-effective and flexible access to scalable compute and storage
resources necessary for today's cloud computing needs. A typical datacenter is made up of …

Reducing internet latency: A survey of techniques and their merits

B Briscoe, A Brunstrom, A Petlund… - … Surveys & Tutorials, 2014 - ieeexplore.ieee.org
Latency is increasingly becoming a performance bottleneck for Internet Protocol (IP)
networks, but historically, networks have been designed with aims of maximizing throughput …

Nightcore: efficient and scalable serverless computing for latency-sensitive, interactive microservices

Z Jia, E Witchel - Proceedings of the 26th ACM International Conference …, 2021 - dl.acm.org
The microservice architecture is a popular software engineering approach for building
flexible, large-scale online services. Serverless functions, or function as a service (FaaS) …

The express data path: Fast programmable packet processing in the operating system kernel

T Høiland-Jørgensen, JD Brouer, D Borkmann… - Proceedings of the 14th …, 2018 - dl.acm.org
Programmable packet processing is increasingly implemented using kernel bypass
techniques, where a userspace application takes complete control of the networking …

Shenango: Achieving high {CPU} efficiency for latency-sensitive datacenter workloads

A Ousterhout, J Fried, J Behrens, A Belay… - … USENIX Symposium on …, 2019 - usenix.org
Datacenter applications demand microsecond-scale tail latencies and high request rates
from operating systems, and most applications handle loads that have high variance over …

Homa: A receiver-driven low-latency transport protocol using network priorities

B Montazeri, Y Li, M Alizadeh… - Proceedings of the 2018 …, 2018 - dl.acm.org
Homa is a new transport protocol for datacenter networks. It provides exceptionally low
latency, especially for workloads with a high volume of very short messages, and it also …

Caladan: Mitigating interference at microsecond timescales

J Fried, Z Ruan, A Ousterhout, A Belay - 14th USENIX Symposium on …, 2020 - usenix.org
The conventional wisdom is that CPU resources such as cores, caches, and memory
bandwidth must be partitioned to achieve performance isolation between tasks. Both the …

Understanding host network stack overheads

Q Cai, S Chaudhary, M Vuppalapati, J Hwang… - Proceedings of the …, 2021 - dl.acm.org
Traditional end-host network stacks are struggling to keep up with rapidly increasing
datacenter access link bandwidths due to their unsustainable CPU overheads. Motivated by …

Congestion control for large-scale RDMA deployments

Y Zhu, H Eran, D Firestone, C Guo… - ACM SIGCOMM …, 2015 - dl.acm.org
Modern datacenter applications demand high throughput (40Gbps) and ultra-low latency (<
10 μs per hop) from the network, with low CPU overhead. Standard TCP/IP stacks cannot …

Datacenter {RPCs} can be general and fast

A Kalia, M Kaminsky, D Andersen - 16th USENIX Symposium on …, 2019 - usenix.org
It is commonly believed that datacenter networking software must sacrifice generality to
attain high performance. The popularity of specialized distributed systems designed …