Proceedings of the 6th International Conference on Supercomputing - ICS '92 1992
DOI: 10.1145/143369.143408
|View full text |Cite
|
Sign up to set email alerts
|

Exploiting heterogeneous parallelism on a multithreaded multiprocessor

Abstract: This paper describes an integrated architecture, compiler, runtime, and operating system solntion to exploiting heteLogeneons parallelism, The archltec I ale is (1 Ijlpc]iilcd mu!tithreaded multiprocessor. enabli]ls the exec II tie]] of J,ery fIHe (multiple operations within an ]nstructton ) to very coarse (multiple jobs) parallel activities. The compiler and IIIntIIILe focus on managing parallelism within a job, while the opelating systemfocuses on managing parallelism across jobs. By considering the entire s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
43
0

Year Published

1998
1998
2014
2014

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 57 publications
(43 citation statements)
references
References 12 publications
0
43
0
Order By: Relevance
“…For both the Alignment and SparseLU benchmarks, BOTS provides two different source files: one in which computation starts with a single initial task and another in which tasks are generated in a loop. 1 Other BOTS benchmarks are not presented here: UTS and FFT use of very fine-grained tasks without cutoffs, yielding poor performance on all run times, and floorplan raises compilation issues in ROSE.…”
Section: Discussionmentioning
confidence: 99%
“…For both the Alignment and SparseLU benchmarks, BOTS provides two different source files: one in which computation starts with a single initial task and another in which tasks are generated in a loop. 1 Other BOTS benchmarks are not presented here: UTS and FFT use of very fine-grained tasks without cutoffs, yielding poor performance on all run times, and floorplan raises compilation issues in ROSE.…”
Section: Discussionmentioning
confidence: 99%
“…In the early 90s the Tera Corporation, starting from the experience acquired with the Horizon machine, built the Tera MTA (Alverson et al, 1990(Alverson et al, , 1992. The MTA design consisted of 256 processors sharing 64 GB of memory organised as a distributed NUMA architecture.…”
Section: Multithreaded Architecturesmentioning
confidence: 99%
“…It is not possible to co-locate data and computation on the same node and data caches cannot be efficiently exploited. Multithreaded architectures, such as the Tera MTA (Alverson et al, 1990(Alverson et al, , 1992Snavely et al, 1998), are designed to mask memory long latency by rapidly switching between concurrent threads; hence, they are ideally suited for irregular applications.…”
Section: Introductionmentioning
confidence: 99%
“…In the early 90's the Tera Corporation, building upon the experience acquired with the Horizon machine, built the Tera MTA [2,1]. The MTA design consisted of 256 processors sharing 64 GB of memory organized as a distributed NUMA architecture.…”
Section: Multithreaded Architecturesmentioning
confidence: 99%