IEEE/ACM International Conference on Computer Aided Design. ICCAD - 2000. IEEE/ACM Digest of Technical Papers (Cat. No.00CH3714
DOI: 10.1109/iccad.2000.896510
|View full text |Cite
|
Sign up to set email alerts
|

MIST: an algorithm for memory miss traffic management

Abstract: Cache misses represent a major bottleneck in embedded systems performance. Traditionally, compilers optimistically treated all memory accesses as cache hits, relying on the memory controller to account for longer miss delays. Howevel; the memory controllerhas only a local view of the program, and is not able to eficiently hide the latency of these memory operations. Our compiler technique actively manages cache misses, and performs global miss trafic optimizations, to better hide the latency of the memory oper… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 14 publications
(7 citation statements)
references
References 21 publications
0
7
0
Order By: Relevance
“…As described earlier, we have already used this memory-aware ADL to generate a compiler [Grun et al 2000a] and manage the memory miss traffic [Grun et al 2000b], resulting in significantly improved performance. We performed comparative studies with the MULTI integrated development environment (IDE) version 3.5 from Green Hills Software Inc. [2003] for the MIPS R4000 processor.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…As described earlier, we have already used this memory-aware ADL to generate a compiler [Grun et al 2000a] and manage the memory miss traffic [Grun et al 2000b], resulting in significantly improved performance. We performed comparative studies with the MULTI integrated development environment (IDE) version 3.5 from Green Hills Software Inc. [2003] for the MIPS R4000 processor.…”
Section: Methodsmentioning
confidence: 99%
“…Section 4 shows an example of performance improvement due to this detailed memory subsystem timing information [Grun et al 2000a]. Such aggressive optimizations in the presence of efficient memory access modes (e.g., page/burst modes) and cache hierarchies [Grun et al 2000b] are only possible due to the explicit representation of the detailed memory architecture. We generate memory simulator (shown shaded in Figure 1) that is integrated into the SIMPRESS ] simulator, allowing for detailed feedback on the memory subsystem architecture and its match to the target applications.…”
Section: Our Approachmentioning
confidence: 99%
“…The references rX , rZ , rY and rW are not part of a reuse pair, since memory lines W, X, Y and Z are accessed only once in the stream. Reuse pair r 1 A , r 2 A has reuse distance 4, while the reuse pair r 2 A , r 3 A has reuse distance 0. The forward reuse distance of r 1 A is 4, its backward reuse distance is ∞.…”
Section: Reuse Distancementioning
confidence: 99%
“…Then, the technique was extended so that it can be well incorporated into loop tiling. More recently, Grun et al presented a new compiler optimization which uses accurate memory access timing information for both cache hits and misses, and schedules instructions so that memory accesses are efficiently overlapped [23].…”
Section: B Memory-aware Code Generationmentioning
confidence: 99%