2012
DOI: 10.21236/ada604494
|View full text |Cite
|
Sign up to set email alerts
|

Tensor Decompositions for Learning Latent Variable Models

Abstract: This work considers a computationally and statistically efficient parameter estimation method for a wide class of latent variable models-including Gaussian mixture models, hidden Markov models, and latent Dirichlet allocation-which exploits a certain tensor structure in their loworder observable moments (typically, of second-and third-order). Specifically, parameter estimation is reduced to the problem of extracting a certain (orthogonal) decomposition of a symmetric tensor derived from the moments; this decom… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

9
1,340
0
3

Year Published

2014
2014
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 484 publications
(1,352 citation statements)
references
References 70 publications
(99 reference statements)
9
1,340
0
3
Order By: Relevance
“…Recently a number of researchers have developed provably correct algorithms for parameter estimation in latent variable models such as hidden Markov models, topic models, directed graphical models with latent variables, and so on (Hsu et al, 2009;Bailly et al, 2010;Siddiqi et al, 2010;Parikh et al, 2011;Balle et al, 2011;Arora et al, 2013;Dhillon et al, 2012;Anandkumar et al, 2012;Arora et al, 2012;Arora et al, 2013). Many of these algorithms have their roots in spectral methods such as canonical correlation analysis (CCA) (Hotelling, 1936), or higher-order tensor decompositions.…”
Section: Related Workmentioning
confidence: 99%
“…Recently a number of researchers have developed provably correct algorithms for parameter estimation in latent variable models such as hidden Markov models, topic models, directed graphical models with latent variables, and so on (Hsu et al, 2009;Bailly et al, 2010;Siddiqi et al, 2010;Parikh et al, 2011;Balle et al, 2011;Arora et al, 2013;Dhillon et al, 2012;Anandkumar et al, 2012;Arora et al, 2012;Arora et al, 2013). Many of these algorithms have their roots in spectral methods such as canonical correlation analysis (CCA) (Hotelling, 1936), or higher-order tensor decompositions.…”
Section: Related Workmentioning
confidence: 99%
“…The definition of the SEY decomposition is not vacuous; it is shown in this paper that orthogonally diagonalizable tensors [18,58] satisfy the above conditions. Such tensors appear in several applications and have been extensively studied [3,6,14,18,46,58]; nevertheless, it does not appear to be known that this decomposition is optimal in the above sense. In fact, we prove that orthogonal diagonalizability is not a necessary condition for an SEY decomposition; a new class of optimally truncatable tensors is revealed in this paper.…”
mentioning
confidence: 99%
“…In further works, we would like to investigate how using NNSpectral to initialize an EM algorithm and how adding constraints in NNSpectral to ensure the convergence of the series. In addition, relations with NMF could be further exploit to produce tensor [1] or kernel-based algorithms.…”
Section: Resultsmentioning
confidence: 99%
“…A detailed description of each problem can be found in [17]. Table 1 results among the following MoM-based algorithms : CO [6] using strings, Tensor [1] using strings and Spectral using strings and substrings. A description and comparison of these algorithms can be found in [5].…”
Section: Numerical Experimentsmentioning
confidence: 99%
See 1 more Smart Citation