Sequential recommendation by reprogramming pretrained transformer

被引:0
作者
Tang, Min [1 ]
Cui, Shujie [2 ]
Jin, Zhe [3 ]
Liang, Shiuan-ni [1 ]
Li, Chenliang [4 ]
Zou, Lixin [4 ]
机构
[1] Monash Univ, Sch Engn, Bandar Sunway 47500, Malaysia
[2] Monash Univ, Sch Informat Technol, Clayton, Vic 3800, Australia
[3] Anhui Univ, Sch Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Generative pretrained transformer; Few-shot learning;
D O I
10.1016/j.ipm.2024.103938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the success of Pre-trained language models (PLMs), numerous sequential recommenders attempted to replicate its achievements by employing PLMs' efficient architectures for building large models and using self-supervised learning for broadening training data. Despite their success, there is curiosity about developing a large-scale sequential recommender system since existing methods either build models within a single dataset or utilize text as an intermediary for alignment across different datasets. However, due to the sparsity of user- item interactions, unalignment between different datasets, and lack of global information in the sequential recommendation, directly pre-training a large foundation model may not be feasible. Towards this end, we propose the RecPPT that firstly employs the GPT-2 to model historical sequence by training the input item embedding and the output layer from scratch, which avoids training a large model on the sparse user-item interactions. Additionally, to alleviate the burden of unalignment, the RecPPT is equipped with a reprogramming module to reprogram the target embedding to existing well-trained proto-embeddings. Furthermore, RecPPT integrates global information into sequences by initializing the item embedding using an SVD-based initializer. Extensive experiments over four datasets demonstrated the RecPPT achieved an average improvement of 6.5% on NDCG@5, 6.2% on NDCG@10, 6.1% on Recall@5, and 5.4% on Recall@10 compared to the baselines. Particularly in few-shot scenarios, the significant improvements in NDCG@10 confirm the superiority of the proposed method.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation
    Ji, Wei
    Liu, Xiangyan
    Zhang, An
    Wei, Yinwei
    Ni, Yongxin
    Wang, Xiang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 955 - 965
  • [22] Knowledge-Enhanced Conversational Recommendation via Transformer-Based Sequential Modeling
    Zou, Jie
    Sun, Aixin
    Long, Cheng
    Kanoulas, Evangelos
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (06)
  • [23] BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer
    Sun, Fei
    Liu, Jun
    Wu, Jian
    Pei, Changhua
    Lin, Xiao
    Ou, Wenwu
    Jiang, Peng
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 1441 - 1450
  • [24] Is News Recommendation a Sequential Recommendation Task?
    Wu, Chuhan
    Wu, Fangzhao
    Qi, Tao
    Li, Chenliang
    Huang, Yongfeng
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2382 - 2386
  • [25] Attention Mixture based Multi-scale Transformer for Multi-behavior Sequential Recommendation
    Li, Tianyang
    Yan, Hongbin
    Jiang, Yuxin
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 2418 - 2423
  • [26] MGT: Multi-Granularity Transformer leveraging multi-level relation for sequential recommendation
    Zhang, Yihu
    Yang, Bo
    Mao, Runze
    Li, Qing
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [27] Reciprocal Sequential Recommendation
    Zheng, Bowen
    Hou, Yupeng
    Zhao, Wayne Xin
    Song, Yang
    Zhu, Hengshu
    PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 89 - 100
  • [28] A Future of Smarter Digital Health Empowered by Generative Pretrained Transformer
    Miao, Hongyu
    Li, Chengdong
    Wang, Jing
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2023, 25
  • [29] Augmenting Sequential Recommendation with Pseudo-Prior Items via Reversely Pre-training Transformer
    Liu, Zhiwei
    Fan, Ziwei
    Wang, Yu
    Yu, Philip S.
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 1608 - 1612
  • [30] Diffusion Augmentation for Sequential Recommendation
    Liu, Qidong
    Yan, Fan
    Zhao, Xiangyu
    Du, Zhaocheng
    Guo, Huifeng
    Tang, Ruiming
    Tian, Feng
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 1576 - 1586