Sequential recommendation by reprogramming pretrained transformer

被引:2
作者
Tang, Min [1 ]
Cui, Shujie [2 ]
Jin, Zhe [3 ]
Liang, Shiuan-ni [1 ]
Li, Chenliang [4 ]
Zou, Lixin [4 ]
机构
[1] Monash Univ, Sch Engn, Bandar Sunway 47500, Malaysia
[2] Monash Univ, Sch Informat Technol, Clayton, Vic 3800, Australia
[3] Anhui Univ, Sch Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Generative pretrained transformer; Few-shot learning;
D O I
10.1016/j.ipm.2024.103938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the success of Pre-trained language models (PLMs), numerous sequential recommenders attempted to replicate its achievements by employing PLMs' efficient architectures for building large models and using self-supervised learning for broadening training data. Despite their success, there is curiosity about developing a large-scale sequential recommender system since existing methods either build models within a single dataset or utilize text as an intermediary for alignment across different datasets. However, due to the sparsity of user- item interactions, unalignment between different datasets, and lack of global information in the sequential recommendation, directly pre-training a large foundation model may not be feasible. Towards this end, we propose the RecPPT that firstly employs the GPT-2 to model historical sequence by training the input item embedding and the output layer from scratch, which avoids training a large model on the sparse user-item interactions. Additionally, to alleviate the burden of unalignment, the RecPPT is equipped with a reprogramming module to reprogram the target embedding to existing well-trained proto-embeddings. Furthermore, RecPPT integrates global information into sequences by initializing the item embedding using an SVD-based initializer. Extensive experiments over four datasets demonstrated the RecPPT achieved an average improvement of 6.5% on NDCG@5, 6.2% on NDCG@10, 6.1% on Recall@5, and 5.4% on Recall@10 compared to the baselines. Particularly in few-shot scenarios, the significant improvements in NDCG@10 confirm the superiority of the proposed method.
引用
收藏
页数:15
相关论文
共 50 条
[21]   Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation [J].
Ji, Wei ;
Liu, Xiangyan ;
Zhang, An ;
Wei, Yinwei ;
Ni, Yongxin ;
Wang, Xiang .
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, :955-965
[22]   Knowledge-Enhanced Conversational Recommendation via Transformer-Based Sequential Modeling [J].
Zou, Jie ;
Sun, Aixin ;
Long, Cheng ;
Kanoulas, Evangelos .
ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (06)
[23]   HyTiFRec: Hybrid Time-Frequency Dual-Branch Transformer for Sequential Recommendation [J].
Qiu, Dawei ;
Wu, Peng ;
Zhang, Xiaoming ;
Xu, Renjie .
CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 83 (02) :1753-1769
[24]   BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer [J].
Sun, Fei ;
Liu, Jun ;
Wu, Jian ;
Pei, Changhua ;
Lin, Xiao ;
Ou, Wenwu ;
Jiang, Peng .
PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, :1441-1450
[25]   Is News Recommendation a Sequential Recommendation Task? [J].
Wu, Chuhan ;
Wu, Fangzhao ;
Qi, Tao ;
Li, Chenliang ;
Huang, Yongfeng .
PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, :2382-2386
[26]   Attention Mixture based Multi-scale Transformer for Multi-behavior Sequential Recommendation [J].
Li, Tianyang ;
Yan, Hongbin ;
Jiang, Yuxin .
PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, :2418-2423
[27]   MGT: Multi-Granularity Transformer leveraging multi-level relation for sequential recommendation [J].
Zhang, Yihu ;
Yang, Bo ;
Mao, Runze ;
Li, Qing .
EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
[28]   Reciprocal Sequential Recommendation [J].
Zheng, Bowen ;
Hou, Yupeng ;
Zhao, Wayne Xin ;
Song, Yang ;
Zhu, Hengshu .
PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, :89-100
[29]   A Future of Smarter Digital Health Empowered by Generative Pretrained Transformer [J].
Miao, Hongyu ;
Li, Chengdong ;
Wang, Jing .
JOURNAL OF MEDICAL INTERNET RESEARCH, 2023, 25
[30]   Augmenting Sequential Recommendation with Pseudo-Prior Items via Reversely Pre-training Transformer [J].
Liu, Zhiwei ;
Fan, Ziwei ;
Wang, Yu ;
Yu, Philip S. .
SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, :1608-1612