Datacenter traffic control: Understanding techniques and tradeoffs
M Noormohammadpour… - … Surveys & Tutorials, 2017 - ieeexplore.ieee.org
Datacenters provide cost-effective and flexible access to scalable compute and storage
resources necessary for today's cloud computing needs. A typical datacenter is made up of …
resources necessary for today's cloud computing needs. A typical datacenter is made up of …
Reducing internet latency: A survey of techniques and their merits
Latency is increasingly becoming a performance bottleneck for Internet Protocol (IP)
networks, but historically, networks have been designed with aims of maximizing throughput …
networks, but historically, networks have been designed with aims of maximizing throughput …
Nightcore: efficient and scalable serverless computing for latency-sensitive, interactive microservices
The microservice architecture is a popular software engineering approach for building
flexible, large-scale online services. Serverless functions, or function as a service (FaaS) …
flexible, large-scale online services. Serverless functions, or function as a service (FaaS) …
The express data path: Fast programmable packet processing in the operating system kernel
T Høiland-Jørgensen, JD Brouer, D Borkmann… - Proceedings of the 14th …, 2018 - dl.acm.org
Programmable packet processing is increasingly implemented using kernel bypass
techniques, where a userspace application takes complete control of the networking …
techniques, where a userspace application takes complete control of the networking …
Shenango: Achieving high {CPU} efficiency for latency-sensitive datacenter workloads
Datacenter applications demand microsecond-scale tail latencies and high request rates
from operating systems, and most applications handle loads that have high variance over …
from operating systems, and most applications handle loads that have high variance over …
Homa: A receiver-driven low-latency transport protocol using network priorities
Homa is a new transport protocol for datacenter networks. It provides exceptionally low
latency, especially for workloads with a high volume of very short messages, and it also …
latency, especially for workloads with a high volume of very short messages, and it also …
Caladan: Mitigating interference at microsecond timescales
The conventional wisdom is that CPU resources such as cores, caches, and memory
bandwidth must be partitioned to achieve performance isolation between tasks. Both the …
bandwidth must be partitioned to achieve performance isolation between tasks. Both the …
Understanding host network stack overheads
Traditional end-host network stacks are struggling to keep up with rapidly increasing
datacenter access link bandwidths due to their unsustainable CPU overheads. Motivated by …
datacenter access link bandwidths due to their unsustainable CPU overheads. Motivated by …
Congestion control for large-scale RDMA deployments
Modern datacenter applications demand high throughput (40Gbps) and ultra-low latency (<
10 μs per hop) from the network, with low CPU overhead. Standard TCP/IP stacks cannot …
10 μs per hop) from the network, with low CPU overhead. Standard TCP/IP stacks cannot …
Datacenter {RPCs} can be general and fast
It is commonly believed that datacenter networking software must sacrifice generality to
attain high performance. The popularity of specialized distributed systems designed …
attain high performance. The popularity of specialized distributed systems designed …