S3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization

被引:557
作者
Zhou, Kun [1 ]
Wang, Hui [1 ]
Zhao, Wayne Xin [2 ,3 ]
Zhu, Yutao [5 ]
Wang, Sirui [4 ]
Zhang, Fuzheng [4 ]
Wang, Zhongyuan [4 ]
Wen, Ji-Rong [2 ,3 ]
机构
[1] Renmin Univ China, Sch Informat, Beijing, Peoples R China
[2] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
[3] Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
[4] Meituan Dianping Grp, Beijing, Peoples R China
[5] Univ Montreal, Montreal, PQ, Canada
来源
CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT | 2020年
基金
中国国家自然科学基金;
关键词
Self-Supervised Learning; Sequential Recommendation; Mutual Information Maximization;
D O I
10.1145/3340531.3411954
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, significant progress has been made in sequential recommendation with deep learning. Existing neural sequential recommendation models usually rely on the item prediction loss to learn model parameters or data representations. However, the model trained with this loss is prone to suffer from data sparsity problem. Since it overemphasizes the final performance, the association or fusion between context data and sequence data has not been well captured and utilized for sequential recommendation. To tackle this problem, we propose the model S-3-Rec, which stands for Self-Supervised learning for Sequential Recommendation, based on the self-attentive neural architecture. The main idea of our approach is to utilize the intrinsic data correlation to derive self-supervision signals and enhance the data representations via pre-training methods for improving sequential recommendation. For our task, we devise four auxiliary self-supervised objectives to learn the correlations among attribute, item, subsequence, and sequence by utilizing the mutual information maximization (MIM) principle. MIM provides a unified way to characterize the correlation between different types of data, which is particularly suitable in our scenario. Extensive experiments conducted on six real-world datasets demonstrate the superiority of our proposed method over existing state-of-the-art methods, especially when only limited training data is available. Besides, we extend our self-supervised learning method to other recommendation models, which also improve their performance.
引用
收藏
页码:1893 / 1902
页数:10
相关论文
共 30 条
[1]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[2]  
Gutmann MU, 2012, J MACH LEARN RES, V13, P307
[3]  
Hidasi B, 2015, P 4 INT C LEARN REPR, DOI [10.48550/arXiv.1511.06939, DOI 10.48550/ARXIV.1511.06939]
[4]   Parallel Recurrent Neural Network Architectures for Feature-rich Session-based Recommendations [J].
Hidasi, Balazs ;
Quadrana, Massimo ;
Karatzoglou, Alexandros ;
Tikk, Domonkos .
PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, :241-248
[5]  
Hjelm R. D., 2019, ICLR 2019
[6]   Taxonomy-Aware Multi-Hop Reasoning Networks for Sequential Recommendation [J].
Huang, Jin ;
Ren, Zhaochun ;
Zhao, Wayne Xin ;
He, Gaole ;
Wen, Ji-Rong ;
Dong, Daxiang .
PROCEEDINGS OF THE TWELFTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM'19), 2019, :573-581
[7]   Improving Sequential Recommendation with Knowledge-Enhanced Memory Networks [J].
Huang, Jin ;
Zhao, Wayne Xin ;
Dou, Hongjian ;
Wen, Ji-Rong ;
Chang, Edward Y. .
ACM/SIGIR PROCEEDINGS 2018, 2018, :505-514
[8]   Self-Attentive Sequential Recommendation [J].
Kang, Wang-Cheng ;
McAuley, Julian .
2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, :197-206
[9]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001
[10]  
Kong L., 2020, ICLR 2020