DOI: 10.1007/978-3-540-87481-2_24
|View full text |Cite
|
Sign up to set email alerts
|

A Unified View of Matrix Factorization Models

Abstract: Abstract. We present a unified view of matrix factorization that frames the differences among popular methods, such as NMF, Weighted SVD, E-PCA, MMMF, pLSI, pLSI-pHITS, Bregman co-clustering, and many others, in terms of a small number of modeling choices. Many of these approaches can be viewed as minimizing a generalized Bregman divergence, and we show that (i) a straightforward alternating projection algorithm can be applied to almost any model in our unified view; (ii) the Hessian for each projection has sp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
85
0
1

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 107 publications
(86 citation statements)
references
References 35 publications
0
85
0
1
Order By: Relevance
“…However, this is not the only way to deal with this kind of problem. For instance, matrix co-factorization (see, e.g., [28]) and tensor co-factorization can be another paradigm of combining explicit features and hidden features.…”
Section: Summary and Discussionmentioning
confidence: 99%
“…However, this is not the only way to deal with this kind of problem. For instance, matrix co-factorization (see, e.g., [28]) and tensor co-factorization can be another paradigm of combining explicit features and hidden features.…”
Section: Summary and Discussionmentioning
confidence: 99%
“…Similarly, Chapter 8 in Tropp's 2004 PhD thesis [Tro04] explored a number of new regularizers, presenting a range of clustering problems as matrix factorization problems with constraints, and anticipated the k-SVD algorithm [AEB06]. Singh and Gordon [SG08] offered a complete view of the state of the literature on matrix factorization in Table 1 of their 2008 paper, and noted that by changing the loss function and regularizer, one may recover algorithms including PCA, weighted PCA, k-means, k-medians, 1 SVD, probabilistic latent semantic indexing (pLSI), nonnegative matrix factorization with 2 or KL-divergence loss, exponential family PCA, and MMMF. Witten et al introduced the statistics community to sparsity-inducing matrix factorization in a 2009 paper on penalized matrix decomposition, with applications to sparse PCA and canonical correlation analysis [WTH09].…”
Section: Gordon's Generalizedmentioning
confidence: 99%
“…For example, there are variants on alternating minimization (with alternating least squares as a special case) [DLYT76, YDLT76, TYDL77, DL84, DLM09], alternating Newton methods [Gor02,SG08], (stochastic or incremental) gradient descent [KO09, LRS + 10, NRRW11, RRWN11, BRRT12, YYH + 13, RR13], conjugate gradients [RS05,SJ03], expectation minimization (EM) (or "soft-impute") methods [TB99,SJ03,MHT10,HMLZ14], multiplicative updates [LS99], and convex relaxations to semidefinite programs [SRJ04,FHB04,RFP10,FM13].…”
Section: Gordon's Generalizedmentioning
confidence: 99%
“…Different approaches have been unified under a generalized Bregman divergence theory [10]. Matrix factorization has been applied in domains involving time-series data as in music transcription [11], up to EEG processing [12] In comparison we are going to use very fast variations of matrix factorization with very low dimensions, fast learning rate and early stopping.…”
Section: Matrix Factorizationmentioning
confidence: 99%
“…Then the objective function is solved for w and w 0 and the dual form is yield as shown in Equation 10.…”
Section: Support Vector Machinesmentioning
confidence: 99%