A Recommendation Algorithm Based on a Self-supervised Learning Pretrain Transformer

被引:8
作者
Xu, Yu-Hao [1 ]
Wang, Zhen-Hai [1 ]
Wang, Zhi-Ru [1 ]
Fan, Rong [1 ]
Wang, Xing [1 ]
机构
[1] Linyi Univ, Coll Informat Sci & Engn, Linyi, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Self-supervised learning; Pretraining; Sequential recommendation; Transformer;
D O I
10.1007/s11063-022-11053-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Click-through rate (CTR) prediction is crucial research direction for the recommendation, with the goal of predicting the probability that users will click on candidate items. There are studies indicates that users' next click behavior is influenced by their last few clicks; therefore, effective modeling of user behavior sequences to extract user interest representations is an important research topic in CTR prediction. Various networks such as RNN, and transformer, have been applied to implicitly extract user interest in the sequence. However, these studies focus on designing complex network structures for better user behavior modeling, while ignoring the fact that the training methods used in current CTR prediction models may limit the model performance. Specifically, owing to the single training objective of the CTR prediction model, the sequence interest extractor component in the model will not be fully trained due to overemphasis on the final prediction effect during the training process. To solve this issue, this paper proposes a recommendation model based on self-supervised learning to pretrain the transformer (SSPT4Rec), which divides the training into two phases: pretraining and fine-tuning. The transformer is trained by a four-classification pretext task in the pretraining phase, and the weights obtained from the pretraining are used to initialize the transformer in the fine-tuning phase and to fine-tune it in the recommendation task. Extensive experiments on four publicly available datasets reveal that the SSPT4Rec method improves the feature extraction capability of transformer as an interest extractor and outperforms the existing model.
引用
收藏
页码:4481 / 4497
页数:17
相关论文
共 40 条
[1]  
Borgeaud S, 2022, PR MACH LEARN RES
[2]   Intent Contrastive Learning for Sequential Recommendation [J].
Chen, Yongjun ;
Liu, Zhiwei ;
Li, Jia ;
McAuley, Julian ;
Xiong, Caiming .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2172-2182
[3]  
Cheng Heng- Tze, 2016, P 1 WORKSH DEEP LEAR, P7
[4]   Scalable Microring-Based Silicon Clos Switch Fabric With Switch-and-Select Stages [J].
Cheng, Qixiang ;
Bahadori, Meisam ;
Hung, Yu-Han ;
Huang, Yishen ;
Abrams, Nathan ;
Bergman, Keren .
IEEE JOURNAL OF SELECTED TOPICS IN QUANTUM ELECTRONICS, 2019, 25 (05)
[5]   MEANTIME: Mixture of Attention Mechanisms with Multi-temporal Embeddings for Sequential Recommendation [J].
Cho, Sung Min ;
Park, Eunhyeok ;
Yoo, Sungjoo .
RECSYS 2020: 14TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2020, :515-520
[6]  
Clark Kevin, 2020, INFORM SYST RES, DOI DOI 10.48550/ARXIV.2003.10555
[7]  
Conneau A, 2019, ADV NEUR IN, V32
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]   Lighter and Better: Low-Rank Decomposed Self-Attention Networks for Next-Item Recommendation [J].
Fan, Xinyan ;
Liu, Zheng ;
Lian, Jianxun ;
Zhao, Wayne Xin ;
Xie, Xing ;
Wen, Ji-Rong .
SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, :1733-1737
[10]   Deep Learning for Sequential Recommendation: Algorithms, Influential Factors, and Evaluations [J].
Fang, Hui ;
Zhang, Danning ;
Shu, Yiheng ;
Guo, Guibing .
ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2020, 39 (01)