2004
DOI: 10.1016/j.jpdc.2004.03.021
|View full text |Cite
|
Sign up to set email alerts
|

Communication lower bounds for distributed-memory matrix multiplication

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

8
299
0

Year Published

2008
2008
2023
2023

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 195 publications
(307 citation statements)
references
References 24 publications
8
299
0
Order By: Relevance
“…We also prove tight lower bounds on the communication costs of rectangular matrix multiplication in all cases. Some of these bounds have appeared previously in [22], and the new bounds use the same techniques (along with those of [2]). As illustrated in Figure 1, the communication costs naturally divide into three cases that we call one large dimension, two large dimensions, and three large dimensions.…”
Section: Contributionsmentioning
confidence: 99%
See 2 more Smart Citations
“…We also prove tight lower bounds on the communication costs of rectangular matrix multiplication in all cases. Some of these bounds have appeared previously in [22], and the new bounds use the same techniques (along with those of [2]). As illustrated in Figure 1, the communication costs naturally divide into three cases that we call one large dimension, two large dimensions, and three large dimensions.…”
Section: Contributionsmentioning
confidence: 99%
“…Following [22], the classical rectangular matrix multiplication algorithm requires mnk scalar multiplications, which may Algorithm 2 CARMA(A,B,C,m,k,n,P ) Input: A is an m × k matrix and B is a k × n matrix Output:…”
Section: B Communication Cost Lower Boundsmentioning
confidence: 99%
See 1 more Smart Citation
“…Our proofs of optimality for communication and synchronization given in this section and the one to follow all derive from lower bounds on the number of communication steps required in distributed algorithms and are direct applications of previous work, particularly of Hong and Kung [23], Aggarwal and Vitter [3], Savage [33,34] and Irony, Toledo and Tiskin [25].…”
Section: Work-limited Algorithmsmentioning
confidence: 99%
“…Our lower bound results for straight-line programs we derive using the approach of Irony, Toledo and Tiskin [25] (and also of [23,33]), while the result for sorting uses an adversarial argument of Aggarwal and Vitter [3]. The bounds will be stated for Multi-BSP but the lower bound arguments for communication hold more generally, for all distributed algorithms with the same hierarchy of memory sizes and costs of communication, even if there is no bulk synchronization.…”
Section: Lower Boundsmentioning
confidence: 99%