2012 39th Annual International Symposium on Computer Architecture (ISCA) 2012
DOI: 10.1109/isca.2012.6237043
|View full text |Cite
|
Sign up to set email alerts
|

Scale-out processors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
33
0

Year Published

2014
2014
2018
2018

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 90 publications
(36 citation statements)
references
References 17 publications
1
33
0
Order By: Relevance
“…A number of studies have attempted to address the overheads of shared cache latency by advocating private cache implementations [9,10,30,12,28]. A number of these studies recognize that private caches can significantly degrade performance compared to shared caches.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…A number of studies have attempted to address the overheads of shared cache latency by advocating private cache implementations [9,10,30,12,28]. A number of these studies recognize that private caches can significantly degrade performance compared to shared caches.…”
Section: Related Workmentioning
confidence: 99%
“…The interconnect latency can be tackled by removing the interconnection network entirely and designing private LLCs or a hybrid of private and shared LLC [9,10,30,12,28]. While hybrid LLCs provide the capacity benefits of a shared cache with the latency benefits of a private cache, they still suffer from the added L2 miss latency when the application working set is larger than the available L2 cache.…”
Section: Introductionmentioning
confidence: 99%
“…However, it is very unlikely that such an adversarial traffic pattern will occur in real workloads. In addition, a recent study on scale-out processors [29] showed that hierarchical and modular memory hierarchy makes optimal use of die area. Recent processors including SPARC T4 [44] and AMD Bulldozer [3] also have hierarchical memory hierarchy.…”
Section: Worst-case Traffic Pattern Analysismentioning
confidence: 99%
“…• A range of new methods to fairly compare the efficiency of server architectures (Section VI) and scale these architectures on demand to meet workload QoS requirements [6], [7]. NanoStreams advances the state of the art in micro-servers in several ways by: (a) adding application-specific but programmable hardware accelerators to micro-servers, as opposed to existing solutions that use elaborate hardware design flows and target a single algorithm [8]; (b) providing general purpose low latency networking to access accelerators in the datacentre, as opposed to custom fabrics [9]; (c) effectively integrating streaming and accelerator-aware programming models into domain specific software stacks, moving one step ahead of ongoing efforts to unify heterogeneous programming models [10]; (d) significantly improving server energy-efficiency of micro-servers via on demand and QoS-aware scale-out and acceleration [6], [7].…”
Section: Introductionmentioning
confidence: 99%