Sequential recommendation by reprogramming pretrained transformer

被引:0
|
作者
Tang, Min [1 ]
Cui, Shujie [2 ]
Jin, Zhe [3 ]
Liang, Shiuan-ni [1 ]
Li, Chenliang [4 ]
Zou, Lixin [4 ]
机构
[1] Monash Univ, Sch Engn, Bandar Sunway 47500, Malaysia
[2] Monash Univ, Sch Informat Technol, Clayton, Vic 3800, Australia
[3] Anhui Univ, Sch Artificial Intelligence, Hefei 230039, Anhui, Peoples R China
[4] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Sequential recommendation; Generative pretrained transformer; Few-shot learning;
D O I
10.1016/j.ipm.2024.103938
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Inspired by the success of Pre-trained language models (PLMs), numerous sequential recommenders attempted to replicate its achievements by employing PLMs' efficient architectures for building large models and using self-supervised learning for broadening training data. Despite their success, there is curiosity about developing a large-scale sequential recommender system since existing methods either build models within a single dataset or utilize text as an intermediary for alignment across different datasets. However, due to the sparsity of user- item interactions, unalignment between different datasets, and lack of global information in the sequential recommendation, directly pre-training a large foundation model may not be feasible. Towards this end, we propose the RecPPT that firstly employs the GPT-2 to model historical sequence by training the input item embedding and the output layer from scratch, which avoids training a large model on the sparse user-item interactions. Additionally, to alleviate the burden of unalignment, the RecPPT is equipped with a reprogramming module to reprogram the target embedding to existing well-trained proto-embeddings. Furthermore, RecPPT integrates global information into sequences by initializing the item embedding using an SVD-based initializer. Extensive experiments over four datasets demonstrated the RecPPT achieved an average improvement of 6.5% on NDCG@5, 6.2% on NDCG@10, 6.1% on Recall@5, and 5.4% on Recall@10 compared to the baselines. Particularly in few-shot scenarios, the significant improvements in NDCG@10 confirm the superiority of the proposed method.
引用
收藏
页数:15
相关论文
共 50 条
  • [11] Multi-Behavior Sequential Recommendation With Temporal Graph Transformer
    Xia, Lianghao
    Huang, Chao
    Xu, Yong
    Pei, Jian
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (06) : 6099 - 6112
  • [12] Personalization Through User Attributes for Transformer-Based Sequential Recommendation
    Fischer, Elisabeth
    Dallmann, Alexander
    Hotho, Andreas
    RECOMMENDER SYSTEMS IN FASHION AND RETAIL, 2023, 981 : 25 - 43
  • [13] Time-Aware Squeeze-Excitation Transformer for Sequential Recommendation
    Chen, Hongwei
    Liu, Luanxuan
    Chen, Zexi
    Li, Xia
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT IX, 2024, 15024 : 121 - 135
  • [14] Multi-scale Interest Dynamic Hierarchical Transformer for sequential recommendation
    Huang, Nana
    Hu, Ruimin
    Xiong, Mingfu
    Peng, Xiaoran
    Ding, Hongwei
    Jia, Xiaodong
    Zhang, Lingkun
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (19) : 16643 - 16654
  • [15] A global contextual enhanced structural-aware transformer for sequential recommendation
    Zhang, Zhu
    Yang, Bo
    Chen, Xingming
    Li, Qing
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [16] Continuous-Time Sequential Recommendation with Temporal Graph Collaborative Transformer
    Fan, Ziwei
    Liu, Zhiwei
    Zhang, Jiawei
    Xiong, Yun
    Zheng, Lei
    Yu, Philip S.
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 433 - 442
  • [17] Multi-Behavior Hypergraph-Enhanced Transformer for Sequential Recommendation
    Yang, Yuhao
    Huang, Chao
    Xia, Lianghao
    Liang, Yuxuan
    Yu, Yanwei
    Li, Chenliang
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 2263 - 2273
  • [18] Multi-scale Interest Dynamic Hierarchical Transformer for sequential recommendation
    Nana Huang
    Ruimin Hu
    Mingfu Xiong
    Xiaoran Peng
    Hongwei Ding
    Xiaodong Jia
    Lingkun Zhang
    Neural Computing and Applications, 2022, 34 : 16643 - 16654
  • [19] Personalized Behavior-Aware Transformer for Multi-Behavior Sequential Recommendation
    Su, Jiajie
    Chen, Chaochao
    Lin, Zibin
    Li, Xi
    Liu, Weiming
    Zheng, Xiaolin
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 6321 - 6331
  • [20] Sequential Recommendation through Graph Neural Networks and Transformer Encoder with Degree Encoding
    Wang, Shuli
    Li, Xuewen
    Kou, Xiaomeng
    Zhang, Jin
    Zheng, Shaojie
    Wang, Jinlong
    Gong, Jibing
    ALGORITHMS, 2021, 14 (09)