Proceedings 1999 IEEE International Conference on Computer Design: VLSI in Computers and Processors (Cat. No.99CB37040)
DOI: 10.1109/iccd.1999.808570
|View full text |Cite
|
Sign up to set email alerts
|

Energy and performance improvements in microprocessor design using a loop cache

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
47
0

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 55 publications
(47 citation statements)
references
References 3 publications
0
47
0
Order By: Relevance
“…Also, Weiyu Tang et al [11] introduce a Decoder Filter Cache in the instruction memory organization in order to reduce the use of the instruction fetch and decode logic by providing directly decoded instructions to the processor. On the other hand, Nikolaos Bellas et al [12] propose a scheme, where the compiler generates code annotations in order to reduce the possibility of a miss in the loop buffer cache. The drawback of this work is the trade-off between the performance degradation and the power savings, which is created by the selection of the basic blocks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Also, Weiyu Tang et al [11] introduce a Decoder Filter Cache in the instruction memory organization in order to reduce the use of the instruction fetch and decode logic by providing directly decoded instructions to the processor. On the other hand, Nikolaos Bellas et al [12] propose a scheme, where the compiler generates code annotations in order to reduce the possibility of a miss in the loop buffer cache. The drawback of this work is the trade-off between the performance degradation and the power savings, which is created by the selection of the basic blocks.…”
Section: Related Workmentioning
confidence: 99%
“…The architectural model that represents it is the central loop buffer architecture for single processor organization. References [5], [6], [7], [8], [9], [10], [11] and [12] are examples of the work done in this set of architectures.…”
Section: Related Workmentioning
confidence: 99%
“…VLIW instruction scheduling has been studied [16][17][18][19] whereas others have considered dynamic voltage scaling techniques [20] and the use of compiler controlled caches for frequently executed code [21]. superscalar processors, most contributions have considered dynamic voltage scaling techniques [20,22].…”
Section: Related Workmentioning
confidence: 99%
“…Dynamic binary translation methods profile in order to store the translation results of frequent code regions, for improved performance as well as power [13], while dynamic optimization methods search for the hottest blocks for runtime recompilation [3]. The approaches in [4][9] use profiling to detect frequent loops to map to a special address region that an architecture would then map to a small low-power loop cache, while the approach in [12] compresses those regions to reduce memory traffic and hence power. The approach in [6] profiles values of variables or subroutine parameters to detect pseudo-constants that can aid a compiler in optimizing for performance, or even for reduced energy [7].…”
Section: Introductionmentioning
confidence: 99%