Proceedings. Second International Symposium on High-Performance Computer Architecture
DOI: 10.1109/hpca.1996.501191
|View full text |Cite
|
Sign up to set email alerts
|

Distributed prefetch-buffer/cache design for high performance memory systems

Abstract: Microprocessor execution speeds are improving at a rate of 50%-80% per year while DRAM access times are improving at a much lower rate of 5%-10% per gear. Computer systems are rapidly approaching the point at which overall system performance is determined not by the speed of the CPU but by the memory system speed. W e present a high performance memory system architecture that overcomes the growing speed disparity between high performance microprocessors and current generation D R A M S . A novel prediction and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
56
0

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 45 publications
(56 citation statements)
references
References 14 publications
0
56
0
Order By: Relevance
“…Although these traces are from the obsolete SPEC92 benchmarks, they are sufficient to warm up the size 1 The traces used in this paper can be found at ftp: //tracebase.nmsu.edu/pub/traces/uni/r2000/ of cache used here, because 1.1-billion references are used, with traces interleaved to create the effect of a multiprogramming workload. Traces of context-switching code and TLB management code are interleaved as appropriate.…”
Section: Inputs and Variationsmentioning
confidence: 99%
See 1 more Smart Citation
“…Although these traces are from the obsolete SPEC92 benchmarks, they are sufficient to warm up the size 1 The traces used in this paper can be found at ftp: //tracebase.nmsu.edu/pub/traces/uni/r2000/ of cache used here, because 1.1-billion references are used, with traces interleaved to create the effect of a multiprogramming workload. Traces of context-switching code and TLB management code are interleaved as appropriate.…”
Section: Inputs and Variationsmentioning
confidence: 99%
“…Prefetch requires loading a cache block before it is requested, either by hardware [7,20] or with compiler support [28]; predictive prefetch attempts to improve accuracy of prefetch for relatively varied memory access patterns [1]. In critical word first, the word containing the reference which caused the miss is fetched first, followed by the rest of the block [13].…”
Section: Alternativesmentioning
confidence: 99%
“…Alexander and Kedem [1] describe a memory-based prefetching scheme that can significantly improve the performance of some applications. They use a prediction table to store up to four possible "next-access" predictions for any given memory address.…”
Section: Related Workmentioning
confidence: 99%
“…Software prefetching [Callahan et al 1991] [Porterfield 1989] [Klaiber and Levy 1991] [Mowry et al 1992] exploits compile-time information to insert prefetch instructions in a program. Correlation-based prefetching [Joseph and Grunwald 1997] [Alexander and Kedem 1996] also relies the address history to predict future references, but they can capture complex access patterns. The prediction accuracy relies on the size of the prediction table and stable access patterns.…”
Section: Related Workmentioning
confidence: 99%