Proceedings of the 47th International Conference on Parallel Processing 2018
DOI: 10.1145/3225058.3225122
|View full text |Cite
|
Sign up to set email alerts
|

Implementing Push-Pull Efficiently in GraphBLAS

Abstract: We factor Beamer's push-pull, also known as direction-optimized breadth-first-search (DOBFS) into 3 separable optimizations, and analyze them for generalizability, asymptotic speedup, and contribution to overall speedup. We demonstrate that masking is critical for high performance and can be generalized to all graph algorithms where the sparsity pattern of the output is known a priori. We show that these graph algorithm optimizations, which together constitute DOBFS, can be neatly and separably described using… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
32
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 34 publications
(32 citation statements)
references
References 23 publications
(29 reference statements)
0
32
0
Order By: Relevance
“…The sequential and parallel versions of this algorithm are deterministic and asymptotically optimal for any ordering of matrix and vector indices. The current state-of-the-art SpMmSpV-BFS approaches are only optimal if the vector indices are unordered [1,25]. It also appears that other recent SpMmSpV methods take O (mn) time overall for BFS because their masking method requires an elementwise multiplication with a dense vector or explicitly testing every vertex in each step [6,25,26].…”
Section: Theorem 1 Bfs Can Be Computed By Xmentioning
confidence: 99%
See 4 more Smart Citations
“…The sequential and parallel versions of this algorithm are deterministic and asymptotically optimal for any ordering of matrix and vector indices. The current state-of-the-art SpMmSpV-BFS approaches are only optimal if the vector indices are unordered [1,25]. It also appears that other recent SpMmSpV methods take O (mn) time overall for BFS because their masking method requires an elementwise multiplication with a dense vector or explicitly testing every vertex in each step [6,25,26].…”
Section: Theorem 1 Bfs Can Be Computed By Xmentioning
confidence: 99%
“…But there is an analysis gap on the asymptotic cost of preventing previous frontier vertices in the BFS from reappearing in the sparse vector. Masking out these frontier nonzeros was analyzed in [25] and it appears to require an elementwise multiplication with a dense masking vector which must be O (n) size to accommodate all vertices. This suggests these SpMmSpV methods with masking take O (mn) time for BFS.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations