2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis 2010
DOI: 10.1109/sc.2010.3
|View full text |Cite
|
Sign up to set email alerts
|

A Block-Oriented Language and Runtime System for Tensor Algebra with Very Large Arrays

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2011
2011
2020
2020

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(17 citation statements)
references
References 19 publications
0
17
0
Order By: Relevance
“…However, while the working set size 2 of a finite element local assembly kernel rarely exceeds a megabyte, the tensors in quantum chemistry computations are typically huge, even reaching petabytes in size. Consequently, the optimized implementation of individual tensor operations on distributed-memory parallel computers is usually the major concern, as is the case with SIAL [48], libtensor [14], and CTF [50]. Similarly, TensorFlow [1] serves to instrument tensor computations for machine learning purposes on large-scale heterogeneous computing systems, so most of their challenges do not overlap with ours.…”
Section: Related Workmentioning
confidence: 99%
“…However, while the working set size 2 of a finite element local assembly kernel rarely exceeds a megabyte, the tensors in quantum chemistry computations are typically huge, even reaching petabytes in size. Consequently, the optimized implementation of individual tensor operations on distributed-memory parallel computers is usually the major concern, as is the case with SIAL [48], libtensor [14], and CTF [50]. Similarly, TensorFlow [1] serves to instrument tensor computations for machine learning purposes on large-scale heterogeneous computing systems, so most of their challenges do not overlap with ours.…”
Section: Related Workmentioning
confidence: 99%
“…This section presents the block tensor, an approach that partitions the modes of a tensor consistently so that the subtensors (“blocks”) have the same dimensionality as the original tensor. The approach, first introduced by Windus and Pople and then used in BlockTensor library and SIAL, is thus a multidimensional generalization of matrix tiling as illustrated in Figure .…”
Section: Data Structures and Algorithmsmentioning
confidence: 99%
“…developed in 1997–1998 following Windus and Pople's ideas. Similar libraries were developed by the NWCHEM (including optimization of tensor expressions), ACES III, and PSI4 teams.…”
Section: Introductionmentioning
confidence: 99%
“…One should note that the fragmentation of the system in our case serves a completely different purpose then it does in fragment‐based local many‐body methods. Namely, there are two reasons why we introduce a formal fragmentation of the system: (1) we can adjust (lower/raise) the quality of the basis set locally (within a fragment) and (2) splitting the full orbital range into contiguous segments naturally partitions all tensors into subtensors (tensor blocks) and this can be exploited efficiently in massively parallel computations (splitting a tensor into tensor blocks, that is, block‐partitioning of tensors, is a higher‐dimensional analog of matrix tiling). Furthermore, since the range of each index (dimension) of a tensor block is associated with a specific local spatial domain (fragment), one can efficiently exploit tensor sparsity and introduce the concept of the tensor block strength, based on some spatial metric, for example, in some analogy with the measures used by Pulay and Szaebo for orbital domains .…”
Section: Theorymentioning
confidence: 99%