2016 6th Workshop on Irregular Applications: Architecture and Algorithms (IA3) 2016
DOI: 10.1109/ia3.2016.010
|View full text |Cite
|
Sign up to set email alerts
|

Optimizing Sparse Tensor Times Matrix on Multi-core and Many-Core Architectures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
27
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
2
1

Relationship

3
4

Authors

Journals

citations
Cited by 23 publications
(27 citation statements)
references
References 20 publications
0
27
0
Order By: Relevance
“…Figure 1. COO format for a general sparse tensor and sCOO format [41] for a semi-sparse tensor. Table 1 presents the operational intensity of each kernel using a cubical third-order tensor, while all the implementations in the benchmark suite support arbitrary tensor orders.…”
Section: Coordinate Format (Coo)mentioning
confidence: 99%
See 1 more Smart Citation
“…Figure 1. COO format for a general sparse tensor and sCOO format [41] for a semi-sparse tensor. Table 1 presents the operational intensity of each kernel using a cubical third-order tensor, while all the implementations in the benchmark suite support arbitrary tensor orders.…”
Section: Coordinate Format (Coo)mentioning
confidence: 99%
“…However, the transformation process brings non-trivial overhead to the execution arXiv:2001.00660v1 [cs.DC] 2 Jan 2020 of a tensor operation. Mitigating this cost has become attractive for researchers in tensor linear algebra and their applications [17,37,41,48,60]. Irregularity in memory access pa erns and in tensor shape makes poor use of memory subsystems and complicates code, especially for sparse data.…”
Section: Introductionmentioning
confidence: 99%
“…is T algorithm directly operates on the input sparse tensor by avoiding tensor transformation. e explanation of Algorithm 5 can be found in the work [85,90].…”
Section: Tmentioning
confidence: 99%
“…This paper expands on the first independent characterization of the Emu Chick prototype [1] by exploring multiple distributed nodes that consist of those nodelets (see Section 2). Our study uses microbenchmarks and small kernels-namely, STREAM, pointer chasing, and sparse matrix-vector multiplication (SpMV)-as proxies that reflect some of the key characteristics of our motivating computations, which come from sparse and irregular applications [4,5]. Indeed, one larger goal of our work beyond this paper is to develop a performance-portable, Emu-compatible API for Georgia Tech's STINGER open-source streaming graph framework [4] and ParTI [6] tensor decomposition algorithms (e.g.…”
Section: Introductionmentioning
confidence: 99%