2015 IEEE International Solid-State Circuits Conference - (ISSCC) Digest of Technical Papers 2015
DOI: 10.1109/isscc.2015.7062931
|View full text |Cite
|
Sign up to set email alerts
|

4.2 A 20nm 32-Core 64MB L3 cache SPARC M7 processor

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 1 publication
0
6
0
Order By: Relevance
“…Adding this to the fact that latency bottlenecks (both due to fetching instructions from the L3 cache, and fetching data from main memory) are still far from eliminated suggests potential for wider SMT: with more threads per core, as seen in some server chips [30]. This case is strengthened by the low memory bandwidth utilization shown earlier -even with more threads per core bandwidth is unlikely to become a bottleneck.…”
Section: Simultaneous Multi-threadingmentioning
confidence: 98%
“…Adding this to the fact that latency bottlenecks (both due to fetching instructions from the L3 cache, and fetching data from main memory) are still far from eliminated suggests potential for wider SMT: with more threads per core, as seen in some server chips [30]. This case is strengthened by the low memory bandwidth utilization shown earlier -even with more threads per core bandwidth is unlikely to become a bottleneck.…”
Section: Simultaneous Multi-threadingmentioning
confidence: 98%
“…Note however, that, even after we account for 2-wide SMT, 75% of collected fleet samples show an IPC value of 1.2 or less (last plot of Figure 14), compared to a theoretical machine width of 4.0. Adding this to the fact that latency bottlenecks (both due to fetching instructions from the L3 cache, and fetching data from main memory) are still far from eliminated suggests potential for wider SMT: with more threads per core, as seen in some server chips [30]. This case is strengthened by the low memory bandwidth utilization shown earlier -even with more threads per core bandwidth is unlikely to become a bottleneck.…”
Section: Simultaneous Multi-threadingmentioning
confidence: 98%
“…For example, the snappy algorithm was designed specifically to achieve higher (de)compression speeds than gzip, sacrificing compression ratios in the process. Its usage might decrease in the presence of sufficiently fast hardware for better-compressing algorithms [30,39].…”
Section: Datacenter Taxmentioning
confidence: 99%
See 1 more Smart Citation
“…One thread is accelerators for data analytics. Accelerator designs have been proposed for query procssing bottlenecks [11,28,35,38,41,50,55,56]. Researchers have explored customized accelerators for common operators in SQL such as join [56], partition [11,55,56], sort [56], aggregation [56], regex search [28,50,56], and index traversal [38].…”
Section: Discussion and Related Workmentioning
confidence: 99%