Attention Calibration for Transformer-based Sequential Recommendation

被引:14
|
作者
Zhou, Peilin [1 ]
Ye, Qichen [2 ]
Xie, Yueqi [1 ]
Gao, Jingqi [3 ]
Wang, Shoujin [4 ]
Kim, Jae Boum [1 ]
You, Chenyu [5 ]
Kim, Sunghun [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Guangzhou, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
[3] Upstage, Salt Lake City, UT USA
[4] Univ Technol Sydney, Sydney, NSW, Australia
[5] Yale Univ, New Haven, CT 06520 USA
来源
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023 | 2023年
关键词
Sequential Recommendation; Attention Mechanism; Transformer;
D O I
10.1145/3583780.3614785
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based sequential recommendation (SR) has been booming in recent years, with the self-attention mechanism as its key component. Self-attention has been widely believed to be able to effectively select those informative and relevant items from a sequence of interacted items for next-item prediction via learning larger attention weights for these items. However, this may not always be true in reality. Our empirical analysis of some representative Transformer-based SR models reveals that it is not uncommon for large attention weights to be assigned to less relevant items, which can result in inaccurate recommendations. Through further in-depth analysis, we find two factors that may contribute to such inaccurate assignment of attention weights: sub-optimal position encoding and noisy input. To this end, in this paper, we aim to address this significant yet challenging gap in existing works. To be specific, we propose a simple yet effective framework called Attention Calibration for Transformer-based Sequential Recommendation (AC-TSR). In AC-TSR, a novel spatial calibrator and adversarial calibrator are designed respectively to directly calibrates those incorrectly assigned attention weights. The former is devised to explicitly capture the spatial relationships (i.e., order and distance) among items for more precise calculation of attention weights. The latter aims to redistribute the attention weights based on each item's contribution to the next-item prediction. AC-TSR is readily adaptable and can be seamlessly integrated into various existing transformerbased SR models. Extensive experimental results on four benchmark real-world datasets demonstrate the superiority of our proposed AC-TSR via significant recommendation performance enhancements. The source code is available at https://github.com/AIM- SE/AC-TSR.
引用
收藏
页码:3595 / 3605
页数:11
相关论文
共 50 条
  • [31] Attention and Convolution Enhanced Memory Network for Sequential Recommendation
    Liu, Jian
    Zhao, Pengpeng
    Liu, Yanchi
    Xu, Jiajie
    Fang, Junhua
    Zhao, Lei
    Sheng, Victor S.
    Cui, Zhiming
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2019), PT II, 2019, 11447 : 333 - 349
  • [32] Sequential Recommendation Based on Multivariate Hawkes Process Embedding With Attention
    Wang, Dongjing
    Zhang, Xin
    Xiang, Zhengzhe
    Yu, Dongjin
    Xu, Guandong
    Deng, Shuiguang
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (11) : 11893 - 11905
  • [33] AdaMCT: Adaptive Mixture of CNN-Transformer for Sequential Recommendation
    Jiang, Juyong
    Zhang, Peiyan
    Luo, Yingtao
    Li, Chaozhuo
    Kim, Jae Boum
    Zhang, Kai
    Wang, Senzhang
    Xie, Xing
    Kim, Sunghun
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 976 - 986
  • [34] Sequential Recommendation via Stochastic Self-Attention
    Fan, Ziwei
    Liu, Zhiwei
    Wang, Yu
    Wang, Alice
    Nazari, Zahra
    Zheng, Lei
    Peng, Hao
    Yu, Philip S.
    PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 2036 - 2047
  • [35] Dual Contrastive Transformer for Hierarchical Preference Modeling in Sequential Recommendation
    Huang, Chengkai
    Wang, Shoujin
    Wang, Xianzhi
    Yao, Lina
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 99 - 109
  • [36] Efficient Open Domain Question Answering With Delayed Attention in Transformer-Based Models
    Siblini, Wissam
    Challal, Mohamed
    Pasqual, Charlotte
    INTERNATIONAL JOURNAL OF DATA WAREHOUSING AND MINING, 2022, 18 (02)
  • [37] Improving Transformer-based Conversational ASR by Inter-Sentential Attention Mechanism
    Wei, Kun
    Guo, Pengcheng
    Jiang, Ning
    INTERSPEECH 2022, 2022, : 3804 - 3808
  • [38] Transformer-based monocular depth estimation with hybrid attention fusion and progressive regression
    Liu, Peng
    Zhang, Zonghua
    Meng, Zhaozong
    Gao, Nan
    NEUROCOMPUTING, 2025, 620
  • [39] Transformer-based multivariate time series anomaly detection using inter-variable attention mechanism
    Kang, Hyeongwon
    Kang, Pilsung
    KNOWLEDGE-BASED SYSTEMS, 2024, 290
  • [40] TMA-Net: A Transformer-Based Multi-Scale Attention Network for Surgical Instrument Segmentation
    Yang, Lei
    Wang, Hongyong
    Gu, Yuge
    Bian, Guibin
    Liu, Yanhong
    Yu, Hongnian
    IEEE TRANSACTIONS ON MEDICAL ROBOTICS AND BIONICS, 2023, 5 (02): : 323 - 334