Lightweight Self-Attentive Sequential Recommendation

被引:48
作者
Li, Yang [1 ]
Chen, Tong [1 ]
Zhang, Peng-Fei [1 ]
Yin, Hongzhi [1 ]
机构
[1] Univ Queensland, Brisbane, Qld, Australia
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021 | 2021年
基金
澳大利亚研究理事会;
关键词
sequential recommendation; lightweight recommender systems; self-attention mechanism;
D O I
10.1145/3459637.3482448
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Modern deep neural networks (DNNs) have greatly facilitated the development of sequential recommender systems by achieving state-of-the-art recommendation performance on various sequential recommendation tasks. Given a sequence of interacted items, existing DNN-based sequential recommenders commonly embed each item into a unique vector to support subsequent computations of the user interest. However, due to the potentially large number of items, the over-parameterised item embedding matrix of a sequential recommender has become a memory bottleneck for efficient deployment in resource-constrained environments, e.g., smartphones and other edge devices. Furthermore, we observe that the widely-used multi-head self-attention, though being effective in modelling sequential dependencies among items, heavily relies on redundant attention units to fully capture both global and local item-item transition patterns within a sequence. In this paper, we introduce a novel lightweight self-attentive network (LSAN) for sequential recommendation. To aggressively compress the original embedding matrix, LSAN leverages the notion of compositional embeddings, where each item embedding is composed by merging a group of selected base embedding vectors derived from substantially smaller embedding matrices. Meanwhile, to account for the intrinsic dynamics of each item, we further propose a temporal context-aware embedding composition scheme. Besides, we develop an innovative twin-attention network that alleviates the redundancy of the traditional multi-head self-attention while retaining full capacity for capturing long- and short-term (i.e., global and local) item dependencies. Comprehensive experiments demonstrate that LSAN significantly advances the accuracy and memory efficiency of existing sequential recommenders.
引用
收藏
页码:967 / 977
页数:11
相关论文
共 55 条
  • [1] [Anonymous], 2018, SIGKDD, DOI DOI 10.1145/3219819.3219869
  • [2] [Anonymous], 2021, P 30 INT JOINT C ART
  • [3] [Anonymous], 2020, IEEE T KNOWLEDGE DAT, DOI DOI 10.1109/ISCAS45731.2020.9181022
  • [4] [Anonymous], 2013, IJCAI
  • [5] Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
  • [6] Chen Ting, 2018, P MACHINE LEARNING R, P854
  • [7] Learning Elastic Embeddings for Customizing On-Device Recommenders
    Chen, Tong
    Yin, Hongzhi
    Zheng, Yujia
    Huang, Zi
    Wang, Yang
    Wang, Meng
    [J]. KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 138 - 147
  • [8] Sequence-Aware Factorization Machines for Temporal Predictive Analytics
    Chen, Tong
    Yin, Hongzhi
    Quoc Viet Hung Nguyen
    Peng, Wen-Chih
    Li, Xue
    Zhou, Xiaofang
    [J]. 2020 IEEE 36TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2020), 2020, : 1405 - 1416
  • [9] AIR: Attentional Intention-Aware Recommender Systems
    Chen, Tong
    Yin, Hongzhi
    Chen, Hongxu
    Yan, Rui
    Quoc Viet Hung Nguyen
    Li, Xue
    [J]. 2019 IEEE 35TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2019), 2019, : 304 - 315
  • [10] Sequential Recommendation with User Memory Networks
    Chen, Xu
    Xu, Hongteng
    Zhang, Yongfeng
    Tang, Jiaxi
    Cao, Yixin
    Qin, Zheng
    Zha, Hongyuan
    [J]. WSDM'18: PROCEEDINGS OF THE ELEVENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2018, : 108 - 116