Sequential recommendation by reprogramming pretrained transformer

被引:2
作者
Tang, Min [1 ]
Cui, Shujie [2 ]
Jin, Zhe [3 ]
Liang, Shiuan-ni [1 ]
Li, Chenliang [4 ]
Zou, Lixin [4 ]
机构
[1] Monash Univ, Sch Engn, Bandar Sunway 47500, Malaysia
[2] Monash Univ, Sch Informat Technol, Clayton, Vic 3800, Australia
[3] Anhui Univ, Sch Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Generative pretrained transformer; Few-shot learning;
D O I
10.1016/j.ipm.2024.103938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the success of Pre-trained language models (PLMs), numerous sequential recommenders attempted to replicate its achievements by employing PLMs' efficient architectures for building large models and using self-supervised learning for broadening training data. Despite their success, there is curiosity about developing a large-scale sequential recommender system since existing methods either build models within a single dataset or utilize text as an intermediary for alignment across different datasets. However, due to the sparsity of user- item interactions, unalignment between different datasets, and lack of global information in the sequential recommendation, directly pre-training a large foundation model may not be feasible. Towards this end, we propose the RecPPT that firstly employs the GPT-2 to model historical sequence by training the input item embedding and the output layer from scratch, which avoids training a large model on the sparse user-item interactions. Additionally, to alleviate the burden of unalignment, the RecPPT is equipped with a reprogramming module to reprogram the target embedding to existing well-trained proto-embeddings. Furthermore, RecPPT integrates global information into sequences by initializing the item embedding using an SVD-based initializer. Extensive experiments over four datasets demonstrated the RecPPT achieved an average improvement of 6.5% on NDCG@5, 6.2% on NDCG@10, 6.1% on Recall@5, and 5.4% on Recall@10 compared to the baselines. Particularly in few-shot scenarios, the significant improvements in NDCG@10 confirm the superiority of the proposed method.
引用
收藏
页数:15
相关论文
共 50 条
[11]   Dual Contrastive Transformer for Hierarchical Preference Modeling in Sequential Recommendation [J].
Huang, Chengkai ;
Wang, Shoujin ;
Wang, Xianzhi ;
Yao, Lina .
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, :99-109
[12]   Time-Aware Squeeze-Excitation Transformer for Sequential Recommendation [J].
Chen, Hongwei ;
Liu, Luanxuan ;
Chen, Zexi ;
Li, Xia .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT IX, 2024, 15024 :121-135
[13]   Multi-scale Interest Dynamic Hierarchical Transformer for sequential recommendation [J].
Nana Huang ;
Ruimin Hu ;
Mingfu Xiong ;
Xiaoran Peng ;
Hongwei Ding ;
Xiaodong Jia ;
Lingkun Zhang .
Neural Computing and Applications, 2022, 34 :16643-16654
[14]   Personalization Through User Attributes for Transformer-Based Sequential Recommendation [J].
Fischer, Elisabeth ;
Dallmann, Alexander ;
Hotho, Andreas .
RECOMMENDER SYSTEMS IN FASHION AND RETAIL, 2023, 981 :25-43
[15]   Multi-scale Interest Dynamic Hierarchical Transformer for sequential recommendation [J].
Huang, Nana ;
Hu, Ruimin ;
Xiong, Mingfu ;
Peng, Xiaoran ;
Ding, Hongwei ;
Jia, Xiaodong ;
Zhang, Lingkun .
NEURAL COMPUTING & APPLICATIONS, 2022, 34 (19) :16643-16654
[16]   A global contextual enhanced structural-aware transformer for sequential recommendation [J].
Zhang, Zhu ;
Yang, Bo ;
Chen, Xingming ;
Li, Qing .
KNOWLEDGE-BASED SYSTEMS, 2024, 304
[17]   Continuous-Time Sequential Recommendation with Temporal Graph Collaborative Transformer [J].
Fan, Ziwei ;
Liu, Zhiwei ;
Zhang, Jiawei ;
Xiong, Yun ;
Zheng, Lei ;
Yu, Philip S. .
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, :433-442
[18]   Multi-Behavior Hypergraph-Enhanced Transformer for Sequential Recommendation [J].
Yang, Yuhao ;
Huang, Chao ;
Xia, Lianghao ;
Liang, Yuxuan ;
Yu, Yanwei ;
Li, Chenliang .
PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, :2263-2273
[19]   Personalized Behavior-Aware Transformer for Multi-Behavior Sequential Recommendation [J].
Su, Jiajie ;
Chen, Chaochao ;
Lin, Zibin ;
Li, Xi ;
Liu, Weiming ;
Zheng, Xiaolin .
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, :6321-6331
[20]   Sequential Recommendation through Graph Neural Networks and Transformer Encoder with Degree Encoding [J].
Wang, Shuli ;
Li, Xuewen ;
Kou, Xiaomeng ;
Zhang, Jin ;
Zheng, Shaojie ;
Wang, Jinlong ;
Gong, Jibing .
ALGORITHMS, 2021, 14 (09)