Content deleted Content added
Changed users, items etc. to text in math equations |
m fixed dashes using a script, cap; fixed 4x invalid ref position using ARA |
||
Line 3:
{{Recommender systems}}
'''Matrix factorization''' is a class of [[collaborative filtering]] algorithms used in [[recommender system]]s<!-- I agree that this is the description but I think the wording should be different. The scientific community as a whole wouldn't commonly know matrix factorization defined in this way -->. Matrix factorization algorithms work by decomposing the user-item interaction [[Matrix (mathematics)|matrix]] into the product of two lower dimensionality rectangular matrices.<ref name="Koren09">{{cite journal |last1=Koren |first1=Yehuda |last2=Bell |first2=Robert |last3=Volinsky |first3=Chris |title=Matrix Factorization Techniques for Recommender Systems |journal=Computer |date=August 2009 |volume=42 |issue=8 |pages=30–37 |doi=10.1109/MC.2009.263|citeseerx=10.1.1.147.8295 |s2cid=58370896 }}</ref> This family of methods became widely known during the [[Netflix prize]] challenge due to its effectiveness as reported by Simon Funk in his 2006 blog post,<ref name="Funkblog">{{cite web |last1=Funk |first1=Simon |title=Netflix Update: Try This at Home |url=http://sifter.org/~simon/journal/20061211.html}}</ref> where he shared his findings with the research community. The prediction results can be improved by assigning different regularization weights to the latent factors based on items' popularity and users' activeness.<ref>{{Cite journal|last1=ChenHung-Hsuan|last2=ChenPu|date=2019-01-09|title=Differentiating Regularization Weights
== Techniques ==
Line 9:
=== Funk MF ===
The original algorithm proposed by Simon Funk in his blog post
The predicted ratings can be computed as <math>\tilde{R}=H W</math>, where <math>\tilde{R} \in \mathbb{R}^{\text{users} \times \text{items}}</math> is the user-item rating matrix, <math>H \in \mathbb{R}^{\text{users} \times \text{latent factors}}</math> contains the user's latent factors and <math>W \in \mathbb{R}^{\text{latent factors} \times \text{items}}</math> the item's latent factors.
Line 16:
<math>\tilde{r}_{ui} = \sum_{f=0}^{\text{n factors} } H_{u,f}W_{f,i}</math>
It is possible to tune the expressive power of the model by changing the number of latent factors. It has been demonstrated
Funk MF was developed as a ''rating prediction'' problem, therefore it uses explicit numerical ratings as user-item interactions.
Line 33:
<math>\tilde{r}_{ui} = \mu + b_i + b_u + \sum_{f=0}^{\text{n factors}} H_{u,f}W_{f,i}</math>
Where <math>\mu</math> refers to the overall average rating over all items and <math>b_i</math> and <math>b_u</math> refers to the observed deviation of the item <math>i</math> and the user <math>u</math> respectively from the average.<ref>{{Cite journal |last1=Koren |first1=Yehuda |last2=Bell |first2=Robert |last3=Volinsky |first3=Chris |date=August 2009 |title=Matrix factorization techniques for recommender systems |url=https://datajobs.com/data-science-repo/Recommender-Systems-[Netflix].pdf |journal=Computer |pages=45}}</ref> SVD++ has however some disadvantages, with the main drawback being that this method is not ''model-based.'' This means that if a new user is added, the algorithm is incapable of modeling it unless the whole model is retrained. Even though the system might have gathered some interactions for that new user, its latent factors are not available and therefore no recommendations can be computed. This is an example of a [[Cold start (recommender systems)|cold-start]] problem, that is the recommender cannot deal efficiently with new users or items and specific strategies should be put in place to handle this disadvantage.<ref name="Kluver14">{{cite book|last1=Kluver |first1=Daniel |title=Proceedings of the 8th ACM Conference on Recommender systems
A possible way to address this cold start problem is to modify SVD++ in order for it to become a ''model-based'' algorithm, therefore allowing to easily manage new items and new users.
Line 45:
=== Asymmetric SVD ===
Asymmetric SVD aims at combining the advantages of SVD++ while being a model based algorithm, therefore being able to consider new users with a few ratings without needing to retrain the whole model. As opposed to the model-based SVD here the user latent factor matrix H is replaced by Q, which learns the user's preferences as function of their ratings.<ref name="Pu13">{{cite book |last1=Pu |first1=Li |title=Proceedings of the 7th ACM conference on Recommender systems
The predicted rating user ''u'' will give to item ''i'' is computed as:
Line 67:
=== Hybrid MF ===
In recent years many other matrix factorization models have been developed to exploit the ever increasing amount and variety of available interaction data and use cases. Hybrid matrix factorization algorithms are capable of merging explicit and implicit interactions
===Deep-
In recent years a number of neural and deep-learning techniques have been proposed, some of which generalize traditional Matrix factorization algorithms via a non-linear neural architecture.<ref>{{cite book |last1=He |first1=Xiangnan |last2=Liao |first2=Lizi |last3=Zhang |first3=Hanwang |last4=Nie |first4=Liqiang |last5=Hu |first5=Xia |last6=Chua |first6=Tat-Seng |title=Proceedings of the 26th International Conference on World Wide Web |chapter=Neural Collaborative Filtering |date=2017 |pages=173–182 |doi=10.1145/3038912.3052569 |chapter-url=https://dl.acm.org/citation.cfm?id=3052569 |accessdate=16 October 2019 |publisher=International World Wide Web Conferences Steering Committee|isbn=9781450349130 |arxiv=1708.05031 |s2cid=13907106 }}</ref>
While deep learning has been applied to many different scenarios: context-aware, sequence-aware, social tagging etc. its real effectiveness when used in a simple [[Collaborative filtering]] scenario has been put into question. Systematic analysis of publications applying deep learning or neural methods to the top-k recommendation problem, published in top conferences (SIGIR, KDD, WWW, RecSys, IJCAI), has shown that on average less than 40% of articles are reproducible, with as little as 14% in some conferences. Overall the studies identify 26 articles, only 12 of them could be reproduced and 11 of them could be outperformed by much older and simpler properly tuned baselines. The articles also highlights a number of potential problems in today's research scholarship and call for improved scientific practices in that area.<ref>{{cite book |last1=Rendle |first1=Steffen |last2=Krichene |first2=Walid |last3=Zhang |first3=Li |last4=Anderson |first4=John |title=Fourteenth ACM Conference on Recommender Systems |chapter=Neural Collaborative Filtering vs. Matrix Factorization Revisited |date=22 September 2020 |pages=240–248 |doi=10.1145/3383313.3412488|arxiv=2005.09683 |isbn=9781450375832 |doi-access=free }}</ref><ref>{{cite journal | last1=Dacrema |last2=Ferrari |title=A Troubling Analysis of Reproducibility and Progress in Recommender Systems Research. |journal=ACM Transactions on Information Systems |pages=39.2 |date=2021 |volume=39 |issue=2 |doi=10.1145/3434185 |arxiv=1911.07698 |s2cid=208138060 }}</ref> Similar issues have been spotted also in sequence-aware recommender systems.<ref>{{cite book |last1=Ludewig |first1=Malte |last2=Mauro |first2=Noemi |last3=Latifi |first3=Sara |last4=Jannach |first4=Dietmar |title=Proceedings of the 13th ACM Conference on Recommender Systems |chapter=Performance comparison of neural and non-neural approaches to session-based recommendation |date=2019 |pages=462–466 |doi=10.1145/3298689.3347041 |publisher=ACM|isbn=9781450362436 |doi-access=free }}</ref>
|