Matrix Factorization In Matlab using Stochastic Gradient Descent
6 views (last 30 days)
Show older comments
I have to factorize Matrix R[m*n] to two low-rank Matrices (U[K*m] and V[K*n]), I do this for predicting missing values of R by U and V.

The problem is, for factorizing R I can't use Matlab factorization methods, so I have to work on objective function which minimizes the sum-of-squared-errors for enhancing factorization accuracy:

details are shown below:

My Question in this post is how to minimize function F in Matlab Using Stochastic Gradient Descent method to decompose R into U and V matrices.
1 Comment
Matt J
on 7 Oct 2013
Since your function is not continuous/differentiable (because I_ij is not), I wonder whether any kind of gradient method applies.
How large are R, U, ad V typically. You might be able to use the genetic algorithm ga() in the Global Optimization Toolbox.
Answers (0)
See Also
Categories
Find more on Get Started with MATLAB in Help Center and File Exchange
Products
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!