Sequential Recommendation via Stochastic Self-Attention

被引:90
|
作者
Fan, Ziwei [1 ,5 ]
Liu, Zhiwei [1 ]
Wang, Yu [1 ]
Wang, Alice [2 ]
Nazari, Zahra [2 ]
Zheng, Lei [3 ]
Peng, Hao [4 ]
Yu, Philip S. [1 ]
机构
[1] Univ Illinois, Dept Comp Sci, Chicago, IL 60680 USA
[2] Spotify, New York, NY USA
[3] Pinterest Inc, Chicago, IL USA
[4] Beihang Univ, Sch Cyber Sci & Technol, Beijing, Peoples R China
[5] Spotify Res, New York, NY USA
来源
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22) | 2022年
关键词
Sequential Recommendation; Transformer; Self-Attention; Uncertainty;
D O I
10.1145/3485447.3512077
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Sequential recommendation models the dynamics of a user's previous behaviors in order to forecast the next item, and has drawn a lot of attention. Transformer-based approaches, which embed items as vectors and use dot-product self-attention to measure the relationship between items, demonstrate superior capabilities among existing sequential methods. However, users' real-world sequential behaviors are uncertain rather than deterministic, posing a significant challenge to present techniques. We further suggest that dot-product-based approaches cannot fully capture collaborative transitivity, which can be derived in item-item transitions inside sequences and is beneficial for cold start items. We further argue that BPR loss has no constraint on positive and sampled negative items, which misleads the optimization. We propose a novel STOchastic Self-Attention (STOSA) to overcome these issues. STOSA, in particular, embeds each item as a stochastic Gaussian distribution, the covariance of which encodes the uncertainty. We devise a novel Wasserstein Self-Attention module to characterize item-item position-wise relationships in sequences, which effectively incorporates uncertainty into model training. Wasserstein attentions also enlighten the collaborative transitivity learning as it satisfies triangle inequality. Moreover, we introduce a novel regularization term to the ranking loss, which assures the dissimilarity between positive and the negative items. Extensive experiments on five real-world benchmark datasets demonstrate the superiority of the proposed model over state-of-the-art baselines, especially on cold start items. The code is available in https://github.com/zfan20/STOSA.
引用
收藏
页码:2036 / 2047
页数:12
相关论文
共 50 条
  • [31] Self-attention Based Collaborative Neural Network for Recommendation
    Ma, Shengchao
    Zhu, Jinghua
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2019, 2019, 11604 : 235 - 246
  • [32] Hashtag Recommendation Using LSTM Networks with Self-Attention
    Shen, Yatian
    Li, Yan
    Sun, Jun
    Ding, Wenke
    Shi, Xianjin
    Zhang, Lei
    Shen, Xiajiong
    He, Jing
    CMC-COMPUTERS MATERIALS & CONTINUA, 2019, 61 (03): : 1261 - 1269
  • [33] Pay Attention to Attention for Sequential Recommendation
    Liu, Yuli
    Liu, Min
    Liu, Xiaojing
    PROCEEDINGS OF THE EIGHTEENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2024, 2024, : 890 - 895
  • [34] Sequential Recommendation Based on Long-Term and Short-Term User Behavior with Self-attention
    Wei, Xing
    Zuo, Xianglin
    Yang, Bo
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2019, PT I, 2019, 11775 : 72 - 83
  • [35] A collaborative filtering recommendation algorithm based on DeepWalk and self-attention
    Guo, Jiaming
    Wen, Hong
    Huang, Weihong
    Yang, Ce
    INTERNATIONAL JOURNAL OF COMPUTATIONAL SCIENCE AND ENGINEERING, 2023, 26 (03) : 296 - 304
  • [36] Stochastic Economic Lot Scheduling via Self-Attention Based Deep Reinforcement Learning
    Song, Wen
    Mi, Nan
    Li, Qiqiang
    Zhuang, Jing
    Cao, Zhiguang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (02) : 1457 - 1468
  • [37] Intention-Centric Learning via Dual Attention for Sequential Recommendation
    Zhang, Zhigao
    Wang, Bin
    Xie, Xinqiang
    IEEE ACCESS, 2024, 12 : 2854 - 2867
  • [38] Personalized News Recommendation with CNN and Multi-Head Self-Attention
    Li, Aibin
    He, Tingnian
    Guo, Yi
    Li, Zhuoran
    Rong, Yixuan
    Liu, Guoqi
    2022 IEEE 13TH ANNUAL UBIQUITOUS COMPUTING, ELECTRONICS & MOBILE COMMUNICATION CONFERENCE (UEMCON), 2022, : 102 - 108
  • [39] Leveraging Knowledge Graph and Self-Attention with Residual Block for Paper Recommendation
    Pang, Xinyue
    Nuo, Minghua
    Cao, Jiamin
    2021 INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE BIG DATA AND INTELLIGENT SYSTEMS (HPBD&IS), 2021, : 196 - 201
  • [40] Design Resources Recommendation Based on Word Vectors and Self-Attention Mechanisms
    Sun Q.
    Deng C.
    Gu Z.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2024, 36 (01): : 63 - 72