The application slowdown model: Quantifying and controlling the impact of inter-application interference at shared caches and main memory
In a multi-core system, interference at shared resources (such as caches and main memory)
slows down applications running on different cores. Accurately estimating the slowdown of …
slows down applications running on different cores. Accurately estimating the slowdown of …
[PDF][PDF] Research problems and opportunities in memory systems
O Mutlu, L Subramanian - Supercomputing frontiers and …, 2014 - superfri.susu.ru
The memory system is a fundamental performance and energy bottleneck in almost all
computing systems. Recent system design, application, and technology trends that require …
computing systems. Recent system design, application, and technology trends that require …
MISE: Providing performance predictability and improving fairness in shared main memory systems
Applications running concurrently on a multicore system interfere with each other at the main
memory. This interference can slow down different applications differently. Accurately …
memory. This interference can slow down different applications differently. Accurately …
Rusty: Runtime interference-aware predictive monitoring for modern multi-tenant systems
Modern micro-service and container-based cloud-native applications have leveraged multi-
tenancy as a first class system design concern. The increasing number of co-located …
tenancy as a first class system design concern. The increasing number of co-located …
Application clustering policies to address system fairness with intel's cache allocation technology
V Selfa, J Sahuquillo, L Eeckhout… - 2017 26th …, 2017 - ieeexplore.ieee.org
Achieving system fairness is a major design concern in current multicore processors.
Unfairness arises due to contention in the shared resources of the system, such as the LLC …
Unfairness arises due to contention in the shared resources of the system, such as the LLC …
Hsm: A hybrid slowdown model for multitasking gpus
Graphics Processing Units (GPUs) are increasingly widely used in the cloud to accelerate
compute-heavy tasks. However, GPU-compute applications stress the GPU architecture in …
compute-heavy tasks. However, GPU-compute applications stress the GPU architecture in …
The main memory system: Challenges and opportunities
The memory system is a fundamental performance and energy bottleneck in almost all
computing systems. Recent system design, application, and technology trends that require …
computing systems. Recent system design, application, and technology trends that require …
Pccs: Processor-centric contention-aware slowdown model for heterogeneous system-on-chips
Many slowdown models have been proposed to characterize memory interference of
workloads co-running on heterogeneous System-on-Chips (SoCs). But they are mostly for …
workloads co-running on heterogeneous System-on-Chips (SoCs). But they are mostly for …
Caliper: Interference estimator for multi-tenant environments sharing architectural resources
We introduce Caliper, a technique for accurately estimating performance interference
occurring in shared servers. Caliper overcomes the limitations of prior approaches by …
occurring in shared servers. Caliper overcomes the limitations of prior approaches by …
A review on shared resource contention in multicores and its mitigating techniques
Chip multiprocessor (CMP) systems have become inevitable to meet high computing
demands. In such systems sharing of resources is imperative for better resource utilisation …
demands. In such systems sharing of resources is imperative for better resource utilisation …