2012 IEEE 12th International Conference on Data Mining 2012
DOI: 10.1109/icdm.2012.168
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Coordinate Descent Approaches to Parallel Matrix Factorization for Recommender Systems

Abstract: Abstract-Matrix factorization, when the matrix has missing values, has become one of the leading techniques for recommender systems. To handle web-scale datasets with millions of users and billions of ratings, scalability becomes an important issue. Alternating Least Squares (ALS) and Stochastic Gradient Descent (SGD) are two popular approaches to compute matrix factorization. There has been a recent flurry of activity to parallelize these algorithms. However, due to the cubic time complexity in the target ran… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
155
0
3

Year Published

2015
2015
2019
2019

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 193 publications
(158 citation statements)
references
References 14 publications
0
155
0
3
Order By: Relevance
“…The early use of CD for MF was in [7], but here we consider the e cient implementation in [29]. The idea is to update one column of W and H at a time.…”
Section: Coordinate Descent (Cd)mentioning
confidence: 99%
See 2 more Smart Citations
“…The early use of CD for MF was in [7], but here we consider the e cient implementation in [29]. The idea is to update one column of W and H at a time.…”
Section: Coordinate Descent (Cd)mentioning
confidence: 99%
“…Thus, the construction of (3.8) costs O(|⌦|k) operations. In [29], the time complexity can be reduced to O(|⌦|) by maintaining the following residual…”
Section: Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…Commonly used methods for (3.2) include stochastic gradient descent (SDG) [10,18], alternative least squares (ALS) [19,11] and coordinate descent (CD) [21]. A variant of (3.2) without the quadratic regularization is solved by a weighted alternating method LMaFit [20].…”
Section: L2 Modelsmentioning
confidence: 99%
“…Efficient computational methods for (1.1) include stochastic gradient descent, alternating least squares, and coordinate descent methods [10,18,21]. We notice that the error term of model (1.1) is L2 squared.…”
Section: Introductionmentioning
confidence: 99%