Contrastive Learning with Bidirectional Transformers for Sequential Recommendation

被引:34
作者
Du, Hanwen [1 ]
Shi, Hui [1 ]
Zhao, Pengpeng [1 ]
Wang, Deqing [2 ]
Sheng, Victor S. [3 ]
Liu, Yanchi [4 ]
Liu, Guanfeng [5 ]
Zhao, Lei [1 ]
机构
[1] Soochow Univ, Suzhou, Jiangsu, Peoples R China
[2] Beihang Univ, Beijing, Peoples R China
[3] Texas Tech Univ, Lubbock, TX 79409 USA
[4] Rutgers State Univ, New Brunswick, NJ USA
[5] Macquarie Univ, Sydney, NSW, Australia
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022 | 2022年
关键词
Sequential Recommendation; Bidirectional Sequential Model; Contrastive Learning;
D O I
10.1145/3511808.3557266
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Contrastive learning with Transformer-based sequence encoder has gained predominance for sequential recommendation. It maximizes the agreements between paired sequence augmentations that share similar semantics. However, existing contrastive learning approaches in sequential recommendation mainly center upon left-to-right unidirectional Transformers as base encoders, which are suboptimal for sequential recommendation because user behaviors may not be a rigid left-to-right sequence. To tackle that, we propose a novel framework named Contrastive learning with Bidirectional Transformers for sequential recommendation (CBiT). Specifically, we first apply the slide window technique for long user sequences in bidirectional Transformers, which allows for a more fine-grained division of user sequences. Then we combine the cloze task mask and the dropout mask to generate high-quality positive samples and perform multi-pair contrastive learning, which demonstrates better performance and adaptability compared with the normal one-pair contrastive learning. Moreover, we introduce a novel dynamic loss reweighting strategy to balance between the cloze task loss and the contrastive loss. Experiment results on three public benchmark datasets show that our model outperforms state-of-the-art models for sequential recommendation. Our code is available at this link: https://github.com/hw-du/CBiT/tree/master.
引用
收藏
页码:396 / 405
页数:10
相关论文
共 50 条
[1]  
Ba J. L., 2016, arXiv, DOI 10.48550/arXiv:1607.06450
[2]  
Bai SJ, 2018, Arxiv, DOI [arXiv:1803.01271, DOI 10.48550/ARXIV.1803.01271]
[3]   Deep Learning for AI [J].
Bengio, Yoshua ;
Lecun, Yann ;
Hinton, Geoffrey .
COMMUNICATIONS OF THE ACM, 2021, 64 (07) :58-65
[4]  
Chen T, 2020, PR MACH LEARN RES, V119
[5]  
Chen XL, 2020, Arxiv, DOI arXiv:2003.04297
[6]   An Empirical Study of Training Self-Supervised Vision Transformers [J].
Chen, Xinlei ;
Xie, Saining ;
He, Kaiming .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9620-9629
[7]  
Chung J, 2014, ARXIV
[8]   Deep Neural Networks for YouTube Recommendations [J].
Covington, Paul ;
Adams, Jay ;
Sargin, Emre .
PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, :191-198
[9]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[10]  
Dosovitskiy A, 2021, ICLR