Attention Calibration for Transformer-based Sequential Recommendation

被引:14
|
作者
Zhou, Peilin [1 ]
Ye, Qichen [2 ]
Xie, Yueqi [1 ]
Gao, Jingqi [3 ]
Wang, Shoujin [4 ]
Kim, Jae Boum [1 ]
You, Chenyu [5 ]
Kim, Sunghun [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Guangzhou, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
[3] Upstage, Salt Lake City, UT USA
[4] Univ Technol Sydney, Sydney, NSW, Australia
[5] Yale Univ, New Haven, CT 06520 USA
来源
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023 | 2023年
关键词
Sequential Recommendation; Attention Mechanism; Transformer;
D O I
10.1145/3583780.3614785
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer-based sequential recommendation (SR) has been booming in recent years, with the self-attention mechanism as its key component. Self-attention has been widely believed to be able to effectively select those informative and relevant items from a sequence of interacted items for next-item prediction via learning larger attention weights for these items. However, this may not always be true in reality. Our empirical analysis of some representative Transformer-based SR models reveals that it is not uncommon for large attention weights to be assigned to less relevant items, which can result in inaccurate recommendations. Through further in-depth analysis, we find two factors that may contribute to such inaccurate assignment of attention weights: sub-optimal position encoding and noisy input. To this end, in this paper, we aim to address this significant yet challenging gap in existing works. To be specific, we propose a simple yet effective framework called Attention Calibration for Transformer-based Sequential Recommendation (AC-TSR). In AC-TSR, a novel spatial calibrator and adversarial calibrator are designed respectively to directly calibrates those incorrectly assigned attention weights. The former is devised to explicitly capture the spatial relationships (i.e., order and distance) among items for more precise calculation of attention weights. The latter aims to redistribute the attention weights based on each item's contribution to the next-item prediction. AC-TSR is readily adaptable and can be seamlessly integrated into various existing transformerbased SR models. Extensive experimental results on four benchmark real-world datasets demonstrate the superiority of our proposed AC-TSR via significant recommendation performance enhancements. The source code is available at https://github.com/AIM- SE/AC-TSR.
引用
收藏
页码:3595 / 3605
页数:11
相关论文
共 50 条
  • [21] Transformer-Based Attention Network for In-Vehicle Intrusion Detection
    Nguyen, Trieu Phong
    Nam, Heungwoo
    Kim, Daehee
    IEEE ACCESS, 2023, 11 : 55389 - 55403
  • [22] TRANSFORMER-BASED TEXT-TO-SPEECH WITH WEIGHTED FORCED ATTENTION
    Okamoto, Takuma
    Toda, Tomoki
    Shiga, Yoshinori
    Kawai, Hisashi
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6729 - 6733
  • [23] Transformer-Based Attention Network for Vehicle Re-Identification
    Lian, Jiawei
    Wang, Dahan
    Zhu, Shunzhi
    Wu, Yun
    Li, Caixia
    ELECTRONICS, 2022, 11 (07)
  • [24] Image captioning using transformer-based double attention network
    Parvin, Hashem
    Naghsh-Nilchi, Ahmad Reza
    Mohammadi, Hossein Mahvash
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 125
  • [25] TransAct: Transformer-based Realtime User Action Model for Recommendation at Pinterest
    Xia, Xue
    Eksombatchai, Pong
    Pancha, Nikil
    Badani, Dhruvil Deven
    Wang, Po-Wei
    Gu, Neng
    Joshi, Saurabh Vishwas
    Farahpour, Nazanin
    Zhang, Zhiyuan
    Zhai, Andrew
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5249 - 5259
  • [26] Attention-based context-aware sequential recommendation model
    Yuan, Weihua
    Wang, Hong
    Yu, Xiaomei
    Liu, Nan
    Li, Zhenghao
    INFORMATION SCIENCES, 2020, 510 : 122 - 134
  • [27] Cascaded Cross Attention for Review -based Sequential Recommendation
    Huang, Bingsen
    Luo, Jinwei
    Du, Weihao
    Pan, Weike
    Ming, Zhong
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 170 - 179
  • [28] Sequential recommendation by reprogramming pretrained transformer
    Tang, Min
    Cui, Shujie
    Jin, Zhe
    Liang, Shiuan-ni
    Li, Chenliang
    Zou, Lixin
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (01)
  • [29] TRANSFORMER-BASED LIP-READING WITH REGULARIZED DROPOUT AND RELAXED ATTENTION
    Li, Zhengyang
    Lohrenz, Timo
    Dunkelberg, Matthias
    Fingscheidt, Tim
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 723 - 730
  • [30] Memory-Augmented Attention Network for Sequential Recommendation
    Hu, Cheng
    He, Peijian
    Sha, Chaofeng
    Niu, Junyu
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2019, 2019, 11881 : 228 - 242