2015
DOI: 10.1137/140980326
|View full text |Cite
|
Sign up to set email alerts
|

On Vector-Kronecker Product Multiplication with Rectangular Factors

Abstract: Abstract. The infinitesimal generator matrix underlying a multidimensional Markov chain can be represented compactly by using sums of Kronecker products of small rectangular matrices. For such compact representations, analysis methods based on vector-Kronecker product multiplication need to be employed. When the factors in the Kronecker product terms are relatively dense, vectorKronecker product multiplication can be performed efficiently by the shuffle algorithm. When the factors are relatively sparse, it may… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 32 publications
(22 reference statements)
0
8
0
Order By: Relevance
“…When the factors are relatively sparse, it may be more efficient to obtain nonzeros of the generator in Kronecker form on the fly and multiply them with corresponding elements of the vector [6]. Recently, the shuffle algorithm has been modified so that relevant elements of the vector are multiplied with submatrices of factors in which zero rows and columns are omitted [8]. This approach is shown to avoid unnecessary floating-point operations (flops) that evaluate to zero during the course of the multiplication and possibly reduces the amount of memory used.…”
Section: Introductionmentioning
confidence: 99%
“…When the factors are relatively sparse, it may be more efficient to obtain nonzeros of the generator in Kronecker form on the fly and multiply them with corresponding elements of the vector [6]. Recently, the shuffle algorithm has been modified so that relevant elements of the vector are multiplied with submatrices of factors in which zero rows and columns are omitted [8]. This approach is shown to avoid unnecessary floating-point operations (flops) that evaluate to zero during the course of the multiplication and possibly reduces the amount of memory used.…”
Section: Introductionmentioning
confidence: 99%
“…This comparison is justified by a considerable gain in both memory efficiency and lower CPU usage when using GTA formalism. Dayar and Orhan work [8] is primarily based on optimizing the execution of the shuffle algorithm in order to improve data locality. The optimization also reduces the number of FLOPS, this is accomplished by focusing the computation on the nonzero values of the matrices and thus trying to avoid FLOPS that use zero rows and columns.…”
Section: Related Workmentioning
confidence: 99%
“…The kronecker product matrix is defined as a block matrix formed with a special multiplication between two matrices [8]. The problem is that given N square matrices A (i) of order n i and a vector x ∈ R 1×L where L = N n=1 n i , the complexity of building this matrix is ( N i=1 n 2 i ).…”
Section: Introductionmentioning
confidence: 99%
“…In practice, the matrices Q k,h are sparse [3] and held in sparse row format since the nonzeros in each of its rows indicate the possible transitions from the state with that row index. The advantage of partitioning the reachable state space is the elimination of unreachable states from the set of rows and columns of the generator to avoid unnecessary computational effort (see, for instance, [2,15]) due to unreachable states and to use vectors not larger than |R| in the analysis. The Kronecker form of the blocks Q (i,j) in Q has been studied before for a number of models [21][22][23].…”
Section: Compact Vectors In Kronecker Settingmentioning
confidence: 99%
“…Starting from an initial solution, the compact vector in HTD format was iteratively multiplied with the uniformized generator matrix of a given CTMC in Kronecker form 1000 times. The same numerical experiment was performed with a solution vector the same size as the reachable state space size using an improved version of the shuffle algorithm [15]. For a fixed truncation error tolerance strategy in the HTD format, the two approaches were compared for memory, time, and accuracy, leading to the preliminary conclusion that compact vectors in HTD format become more memory efficient as the number of dimensions increases.…”
Section: Introductionmentioning
confidence: 99%