Difference between revisions of "SVD++"

From RecSysWiki
Jump to navigation Jump to search
Line 11: Line 11:
  
 
== Efficient SGD Training for SVD++ ==
 
== Efficient SGD Training for SVD++ ==
please refer to http://arxiv.org/abs/1109.2271
+
please refer to http://arxiv.org/abs/1109.2271. Describe efficient SVD++ training in section 4
  
 
== Literature ==
 
== Literature ==

Revision as of 20:45, 23 September 2011

SVD++ refers to the matrix factorization algorithm which makes use of implicit feedback information. In general, implicit feedback can refer to any kinds of users' history information that can help indicate users' preference.

Model Formalization

currently seems that Latex formula is not supported, wait for another solution.

Model Learning

  • SVD++ can be trained using ALS.
  • It's a bit unwise to train a SVD++ style model using stochastic gradient descent due to the size of user feedback information, however, an efficient SGD training algorithm can be used.

Efficient SGD Training for SVD++

please refer to http://arxiv.org/abs/1109.2271. Describe efficient SVD++ training in section 4

Literature

Implementations

  • GraphLab Collaborative Filtering Library has implemented SVD++ for multicore: http://graphlab.org/pmf.html
  • SVDFeature is a toolkit designed for feature-based matrix factorization, can be used to implement SVD++ and it's extensions.