1989
DOI: 10.1145/68182.68207
|View full text |Cite
|
Sign up to set email alerts
|

Available instruction-level parallelism for superscalar and superpipelined machines

Abstract: Su erscalar machines can issue several instructions perIp cyc e. Superpipelined machines can issue only one instruction per cycle, but they have cycle times shorter than the latency of any functional unit. In this paper these two techniques are shown to be roughly equivalent ways of exploiting instruction-level parallelism.A parameterizable code reorganization and simulation system was developed and used to measure instruction-level parallelism for a series of benchmarks. Results of these simulations in the pr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0
2

Year Published

1992
1992
2013
2013

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 117 publications
(13 citation statements)
references
References 13 publications
0
9
0
2
Order By: Relevance
“…To maximize the overall performance, the respective code has to be scheduled in a way to take maximum advantage of the pipelines provided by the architecture [40], [47]. Instruction scheduling is an optimization technique that rearranges the micro-operations executed in a processor's pipeline, attempting to maximize the number of functional units operating in parallel and to minimize the time they spend waiting for each other [30].…”
Section: Feedback Driven Optimizationsmentioning
confidence: 99%
“…To maximize the overall performance, the respective code has to be scheduled in a way to take maximum advantage of the pipelines provided by the architecture [40], [47]. Instruction scheduling is an optimization technique that rearranges the micro-operations executed in a processor's pipeline, attempting to maximize the number of functional units operating in parallel and to minimize the time they spend waiting for each other [30].…”
Section: Feedback Driven Optimizationsmentioning
confidence: 99%
“…First, we defined the instruction set as well as the addressing modes we wanted to support (see Appendix II for details). Next, we started to design the internal structure of the CPU using superscalar and superpipeline concepts [9]. Based on this, we divided the CPU pipeline operation into the following stages: Instruction Fetch (IF), Instruction Dispatch (ID), Instruction Decode (D), Address Generation (AG), Operand Fetch (OF), Execution (EX), and Write Back (WB).…”
Section: Layout Of Architecturementioning
confidence: 99%
“…A second column has been inserted after each line number to indicate which processing element(s) process this RTL line. For instance, the first line ( [1]) initializes the variable q to zero. Since the only use of q is in code allocated to PE3, the initialization of q is allocated to PE3.…”
Section: Code Separationmentioning
confidence: 99%