Differentiating Regularization Weights - A Simple Mechanism to Alleviate Cold Start in Recommender Systems

被引:21
作者
Chen, Hung-Hsuan [1 ]
Chen, Pu [1 ]
机构
[1] Natl Cent Univ, Comp Sci & Informat Engn, 300 Zhongda Rd, Taoyuan 32001, Taiwan
关键词
Recommender systems; matrix factorization; collaborative filtering; SVD; SVD plus; cold start; long tail; INFORMATION OVERLOAD; REGRESSION; SELECTION;
D O I
10.1145/3285954
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Matrix factorization (ME) and its extended methodologies have been studied extensively in the community of recommender systems in the last decade. Essentially, MF attempts to search for low-ranked matrices that can (1) best approximate the known rating scores, and (2) maintain low Frobenius norm for the low-ranked matrices to prevent overfitting. Since the two objectives conflict with each other, the common practice is to assign the relative importance weights as the hyper-parameters to these objectives. The two low-ranked matrices returned by MF are often interpreted as the latent factors of a user and the latent factors of an item that would affect the rating of the user on the item. As a result, it is typical that, in the loss function, we assign a regularization weight lambda(p) on the norms of the latent factors for all users, and another regularization weight lambda(q) on the norms of the latent factors for all the items. We argue that such a methodology probably over-simplifies the scenario. Alternatively, we probably should assign lower constraints to the latent factors associated with the items or users that reveal more information, and set higher constraints to the others. In this article, we systematically study this topic. We found that such a simple technique can improve the prediction results of the MF-based approaches based on several public datasets. Specifically, we applied the proposed methodology on three baseline models - SVD, SVD++, and the NMF models. We found that this technique improves the prediction accuracy for all these baseline models. Perhaps more importantly, this technique better predicts the ratings on the long-tail items, i.e., the items that were rated/viewed/purchased by few users. This suggests that this approach may partially remedy the cold-start issue. The proposed method is very general and can be easily applied on various recommendation models, such as Factorization Machines, Field-aware Factorization Machines, Factorizing Personalized Markov Chains, Prod2Vec, Behavior2Vec, and so on. We release the code for reproducibility. We implemented a Python package that integrates the proposed regularization technique with the SVD, SVD++, and the NMF model. The package can be accessed at https://github.com/ncti-dart/rdf.
引用
收藏
页数:22
相关论文
共 59 条
[51]  
Rendle Steffen, 2010, Proceedings 2010 10th IEEE International Conference on Data Mining (ICDM 2010), P995, DOI 10.1109/ICDM.2010.127
[52]  
Rendle Steffen, 2010, WWW, P811
[53]  
Schein A. I., 2002, Proceedings of SIGIR 2002. Twenty-Fifth Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, P253, DOI 10.1145/564376.564421
[54]  
Srivastava N, 2014, J MACH LEARN RES, V15, P1929
[55]  
Sun M., 2013, ACM International Conference on Web Search and Data Mining (WSDM), P445, DOI [10.1145/2433396.2433451, DOI 10.1145/2433396.2433451]
[57]   Recurrent Recommender Networks [J].
Wu, Chao-Yuan ;
Ahmed, Amr ;
Beutel, Alex ;
Smola, Alexander J. ;
Jing, How .
WSDM'17: PROCEEDINGS OF THE TENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2017, :495-503
[58]  
Zhou K, 2011, PROCEEDINGS OF THE 34TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR'11), P315
[59]   Regularization and variable selection via the elastic net [J].
Zou, H ;
Hastie, T .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2005, 67 :301-320