Proceedings of the 22nd Annual International Symposium on Computer Architecture - ISCA '95 1995
DOI: 10.1145/223982.224447
|View full text |Cite
|
Sign up to set email alerts
|

Streamlining data cache access with fast address calculation

Abstract: For many programs, especially integer codes, untolerated load instruction latencies account for a significant portion of total execution time. In this paper, we present the design and evaluation of a fast address generation mechanism capable of eliminating the delays caused by effective address calculation for many loads and stores.Our approach works by predicting early in the pipeline (part of) the effective address of a memory access and using this predicted address to speculatively access the data cache. If… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
37
0

Year Published

1997
1997
2010
2010

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 41 publications
(37 citation statements)
references
References 12 publications
0
37
0
Order By: Relevance
“…Of course, since prefetching has no architected side effects, no mechanism is needed for verifying the accuracy of the prediction or for recovering from mispredictions. Another example of a technique that speculates on data address is fast address calculation [26,11], which enables early initiation of memory loads by speculatively generating addresses early in the pipeline.…”
Section: Data Speculationmentioning
confidence: 99%
“…Of course, since prefetching has no architected side effects, no mechanism is needed for verifying the accuracy of the prediction or for recovering from mispredictions. Another example of a technique that speculates on data address is fast address calculation [26,11], which enables early initiation of memory loads by speculatively generating addresses early in the pipeline.…”
Section: Data Speculationmentioning
confidence: 99%
“…The equivalent of way-prediction for icaches is often combined with branch prediction [5,9], but because D-caches do not interact with branch prediction, those techniques cannot be used directly. An alternative to prediction is to obtain the correct way-number of the displaced block using the address, which delays initiating cache access to the displaced block, as is the case for statically probed schemes such as column-associative and We examine two handles that can be used to perform way prediction: instruction PC and approximate data address formed by XORing the register value with the instruction offset (proposed in [3], and used in [6]), which may be faster than performing a full add. These two handles represent the two extremes of the trade-off between prediction accuracy and early availability in the pipeline, as shown in Figure 3.…”
Section: Way Predictionmentioning
confidence: 99%
“…XOR-based way prediction, used in the PSA paper [6}, relies on the idea that while a pipeline stage computes the data address by adding the source register value to the instruction offset, the register value can be XORed with the instruction offset to compute an approximate of the address [3] and access a way-prediction table. This scheme exploits the fact that most memory instructions have small enough offsets so that the block address from the XOR approximation is usually same as or at least correlates well with the block address from the actual data address.…”
Section: Xor-based Way-predictionmentioning
confidence: 99%
See 1 more Smart Citation
“…They have been mainly used to reduce the influence of conditional branches [11]. In recent works, prediction techniques have also been applied to predict values or addresses to speculatively issue dependent operations [2] [7][9] [10].…”
Section: Introductionmentioning
confidence: 99%