MISS: Multi-Interest Self-Supervised Learning Framework for Click-Through Rate Prediction

被引:18
作者
Guo, Wei [1 ]
Zhang, Can [2 ]
He, Zhicheng [1 ]
Qin, Jiarui [3 ]
Guo, Huifeng [1 ]
Chen, Bo [1 ]
Tang, Ruiming [1 ]
He, Xiuqiang [1 ]
Zhang, Rui [4 ]
机构
[1] Huawei Noahs Ark Lab, Montreal, PQ, Canada
[2] Natl Univ Singapore, Singapore, Singapore
[3] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[4] Wwwruizhanginfo, Melbourne, Vic, Australia
来源
2022 IEEE 38TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2022) | 2022年
关键词
CTR Prediction; Multi-interest; Self-Supervised Learning;
D O I
10.1109/ICDE53745.2022.00059
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
CTR prediction is essential for modern recommender systems. Ranging from early factorization machines to deep learning based models in recent years, existing CTR methods focus on capturing useful feature interactions or mining important behavior patterns. Despite the effectiveness, we argue that these methods suffer from the risk of label sparsity (i.e., the user-item interactions are highly sparse with respect to the feature space), label noise (i.e., the collected user-item interactions are usually noisy), and the underuse of domain knowledge (i.e., the pairwise correlations between samples). To address these challenging problems, we propose a novel Multi-Interest Self-Supervised learning (MISS) framework which enhances the feature embeddings with interest-level self-supervision signals. With the help of two novel CNN-based multi-interest extractors, self-supervision signals are discovered with full considerations of different interest representations (point-wise and union-wise), interest dependencies (short-range and long-range), and interest correlations (inter-item and intra-item). Based on that, contrastive learning losses are further applied to the augmented views of interest representations, which effectively improves the feature representation learning. Furthermore, our proposed MISS framework can be used as an "plug-in" component with existing CTR prediction models and further boost their performances. Extensive experiments on three large-scale datasets show that MISS significantly outperforms the state-of-the-art models, by up to 13.55% in AUC, and also enjoys good compatibility with representative deep CTR models.
引用
收藏
页码:727 / 740
页数:14
相关论文
共 54 条
[1]   Deep Clustering for Unsupervised Learning of Visual Features [J].
Caron, Mathilde ;
Bojanowski, Piotr ;
Joulin, Armand ;
Douze, Matthijs .
COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 :139-156
[2]   Controllable Multi-Interest Framework for Recommendation [J].
Cen, Yukuo ;
Zhang, Jianwei ;
Zou, Xu ;
Zhou, Chang ;
Yang, Hongxia ;
Tang, Jie .
KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, :2942-2951
[3]   Enhancing Explicit and Implicit Feature Interactions via Information Sharing for Parallel Deep CTR Models [J].
Chen, Bo ;
Wang, Yichao ;
Liu, Zhirong ;
Tang, Ruiming ;
Guo, Wei ;
Zheng, Hongkun ;
Yao, Weiwei ;
Zhang, Muyu ;
He, Xiuqiang .
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, :3757-3766
[4]  
Chen T., 2020, ADV NEURIPS
[5]  
Chen T, 2020, PR MACH LEARN RES, V119
[6]  
Cheng H. T., 2016, P 1 WORKSH DEEP LEAR, P7
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Feng YF, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2301
[9]  
Gidaris S., 2018, INT C LEARN REPR ICL
[10]  
Guo HF, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1725