Attention Calibration for Transformer-based Sequential Recommendation

被引:14
|
作者
Zhou, Peilin [1 ]
Ye, Qichen [2 ]
Xie, Yueqi [1 ]
Gao, Jingqi [3 ]
Wang, Shoujin [4 ]
Kim, Jae Boum [1 ]
You, Chenyu [5 ]
Kim, Sunghun [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Guangzhou, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
[3] Upstage, Salt Lake City, UT USA
[4] Univ Technol Sydney, Sydney, NSW, Australia
[5] Yale Univ, New Haven, CT 06520 USA
来源
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023 | 2023年
关键词
Sequential Recommendation; Attention Mechanism; Transformer;
D O I
10.1145/3583780.3614785
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based sequential recommendation (SR) has been booming in recent years, with the self-attention mechanism as its key component. Self-attention has been widely believed to be able to effectively select those informative and relevant items from a sequence of interacted items for next-item prediction via learning larger attention weights for these items. However, this may not always be true in reality. Our empirical analysis of some representative Transformer-based SR models reveals that it is not uncommon for large attention weights to be assigned to less relevant items, which can result in inaccurate recommendations. Through further in-depth analysis, we find two factors that may contribute to such inaccurate assignment of attention weights: sub-optimal position encoding and noisy input. To this end, in this paper, we aim to address this significant yet challenging gap in existing works. To be specific, we propose a simple yet effective framework called Attention Calibration for Transformer-based Sequential Recommendation (AC-TSR). In AC-TSR, a novel spatial calibrator and adversarial calibrator are designed respectively to directly calibrates those incorrectly assigned attention weights. The former is devised to explicitly capture the spatial relationships (i.e., order and distance) among items for more precise calculation of attention weights. The latter aims to redistribute the attention weights based on each item's contribution to the next-item prediction. AC-TSR is readily adaptable and can be seamlessly integrated into various existing transformerbased SR models. Extensive experimental results on four benchmark real-world datasets demonstrate the superiority of our proposed AC-TSR via significant recommendation performance enhancements. The source code is available at https://github.com/AIM- SE/AC-TSR.
引用
收藏
页码:3595 / 3605
页数:11
相关论文
共 50 条
  • [41] Attention Analysis and Calibration for Transformer in Natural Language Generation
    Lu, Yu
    Zhang, Jiajun
    Zeng, Jiali
    Wu, Shuangzhi
    Zong, Chengqing
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 1927 - 1938
  • [42] Improving scene text image captioning using transformer-based multilevel attention
    Srivastava, Swati
    Sharma, Himanshu
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (03)
  • [43] A novel transformer-based network with attention mechanism for automatic pavement crack detection
    Guo, Feng
    Liu, Jian
    Lv, Chengshun
    Yu, Huayang
    CONSTRUCTION AND BUILDING MATERIALS, 2023, 391
  • [44] Transformer-based multi-attention hybrid networks for skin lesion segmentation
    Dong, Zhiwei
    Li, Jinjiang
    Hua, Zhen
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 244
  • [45] Local-Global Self-Attention for Transformer-Based Object Tracking
    Chen, Langkun
    Gao, Long
    Jiang, Yan
    Li, Yunsong
    He, Gang
    Ning, Jifeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (12) : 12316 - 12329
  • [46] Personalized Dual Transformer Network for sequential recommendation
    Ge, Meiling
    Wang, Chengduan
    Qin, Xueyang
    Dai, Jiangyan
    Huang, Lei
    Qin, Qibing
    Zhang, Wenfeng
    NEUROCOMPUTING, 2025, 622
  • [47] Traffic Transformer: Transformer-based framework for temporal traffic accident prediction
    Al-Thani, Mansoor G.
    Sheng, Ziyu
    Cao, Yuting
    Yang, Yin
    AIMS MATHEMATICS, 2024, 9 (05): : 12610 - 12629
  • [48] Transformer-Based Reconstruction for Fourier Ptychographic Microscopy
    Zhao, Lin
    Zhou, Xuhui
    Lu, Xin
    Tong, Haiping
    Fang, Hui
    IEEE ACCESS, 2023, 11 : 94536 - 94544
  • [49] Annular-Graph Attention Model for Personalized Sequential Recommendation
    Hao, Junmei
    Dun, Yujie
    Zhao, Guoshuai
    Wu, Yuxia
    Qian, Xueming
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 24 : 3381 - 3391
  • [50] Sequential Recommendation Model Based on Temporal Convolution Attention Neural Network
    Du Y.
    Niu J.
    Wang L.
    Yan R.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2022, 35 (05): : 472 - 480