2019 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2019
DOI: 10.1109/hpca.2019.00037
|View full text |Cite
|
Sign up to set email alerts
|

Enhancing Server Efficiency in the Face of Killer Microseconds

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 24 publications
(8 citation statements)
references
References 105 publications
0
6
0
Order By: Relevance
“…in large warehouse-scale computers, and present some optimizations that can help mitigate such system bottlenecks. Mirhosseini et al [116] explore killer microseconds -microsecond-scale "holes" in CPU schedules caused by I/O stalls or idle periods between requests in high throughput microservices that are typical in data centers. They then propose enhancements to server architectures to help mitigate such efects.…”
Section: Pue = Total_power_consumption It_power_consumptionmentioning
confidence: 99%
“…in large warehouse-scale computers, and present some optimizations that can help mitigate such system bottlenecks. Mirhosseini et al [116] explore killer microseconds -microsecond-scale "holes" in CPU schedules caused by I/O stalls or idle periods between requests in high throughput microservices that are typical in data centers. They then propose enhancements to server architectures to help mitigate such efects.…”
Section: Pue = Total_power_consumption It_power_consumptionmentioning
confidence: 99%
“…A further option is to utilize larger caches, such as the private (per-core) L2 caches or the shared L3 caches, to also store state for the additional hardware threads, similar to Duplexity [56]. A fraction of a 512KB private L2 cache can store the state of tens of threads, while a few MB of an L3 cache can support hundreds of threads.…”
Section: The Space Of Hardware Designsmentioning
confidence: 99%
“…The combination of PS scheduling with threadper-request will actually provide superior performance for server workloads with high execution-time variability [46,80]. In addition to RR scheduling, we can introduce hardware support for thread priorities (e.g., threads used for serving time-sensitive interrupts receive more cycles [56]) or even hardware-based (but software-managed) thread queuing, load balancing, priorities, and scheduling [29,52,67]. Hardware support will be needed for fine-grain tracking of threads' resource consumption for cloud billing or software decisions.…”
Section: The Space Of Hardware Designsmentioning
confidence: 99%
“…It supports strict priority-based resource allocation policies but fails to consider resource contention between colocated tasks. In addition, Duplexity [47] uses an aggressive multithreading technique to hide the latency in the order of microseconds.…”
Section: Related Workmentioning
confidence: 99%