Lightweight Self-Attentive Sequential Recommendation

被引:64
作者
Li, Yang [1 ]
Chen, Tong [1 ]
Zhang, Peng-Fei [1 ]
Yin, Hongzhi [1 ]
机构
[1] Univ Queensland, Brisbane, Qld, Australia
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021 | 2021年
基金
澳大利亚研究理事会;
关键词
sequential recommendation; lightweight recommender systems; self-attention mechanism;
D O I
10.1145/3459637.3482448
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Modern deep neural networks (DNNs) have greatly facilitated the development of sequential recommender systems by achieving state-of-the-art recommendation performance on various sequential recommendation tasks. Given a sequence of interacted items, existing DNN-based sequential recommenders commonly embed each item into a unique vector to support subsequent computations of the user interest. However, due to the potentially large number of items, the over-parameterised item embedding matrix of a sequential recommender has become a memory bottleneck for efficient deployment in resource-constrained environments, e.g., smartphones and other edge devices. Furthermore, we observe that the widely-used multi-head self-attention, though being effective in modelling sequential dependencies among items, heavily relies on redundant attention units to fully capture both global and local item-item transition patterns within a sequence. In this paper, we introduce a novel lightweight self-attentive network (LSAN) for sequential recommendation. To aggressively compress the original embedding matrix, LSAN leverages the notion of compositional embeddings, where each item embedding is composed by merging a group of selected base embedding vectors derived from substantially smaller embedding matrices. Meanwhile, to account for the intrinsic dynamics of each item, we further propose a temporal context-aware embedding composition scheme. Besides, we develop an innovative twin-attention network that alleviates the redundancy of the traditional multi-head self-attention while retaining full capacity for capturing long- and short-term (i.e., global and local) item dependencies. Comprehensive experiments demonstrate that LSAN significantly advances the accuracy and memory efficiency of existing sequential recommenders.
引用
收藏
页码:967 / 977
页数:11
相关论文
共 55 条
[1]  
[Anonymous], 2018, SIGKDD, DOI DOI 10.1145/3219819.3219869
[2]  
[Anonymous], 2021, P 30 INT JOINT C ART
[3]  
[Anonymous], 2020, IEEE T KNOWLEDGE DAT, DOI DOI 10.1109/ISCAS45731.2020.9181022
[4]  
Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
[5]  
Chen T, 2018, PR MACH LEARN RES, V80
[6]   Learning Elastic Embeddings for Customizing On-Device Recommenders [J].
Chen, Tong ;
Yin, Hongzhi ;
Zheng, Yujia ;
Huang, Zi ;
Wang, Yang ;
Wang, Meng .
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, :138-147
[7]   Sequence-Aware Factorization Machines for Temporal Predictive Analytics [J].
Chen, Tong ;
Yin, Hongzhi ;
Quoc Viet Hung Nguyen ;
Peng, Wen-Chih ;
Li, Xue ;
Zhou, Xiaofang .
2020 IEEE 36TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2020), 2020, :1405-1416
[8]   AIR: Attentional Intention-Aware Recommender Systems [J].
Chen, Tong ;
Yin, Hongzhi ;
Chen, Hongxu ;
Yan, Rui ;
Quoc Viet Hung Nguyen ;
Li, Xue .
2019 IEEE 35TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2019), 2019, :304-315
[9]   Sequential Recommendation with User Memory Networks [J].
Chen, Xu ;
Xu, Hongteng ;
Zhang, Yongfeng ;
Tang, Jiaxi ;
Cao, Yixin ;
Qin, Zheng ;
Zha, Hongyuan .
WSDM'18: PROCEEDINGS OF THE ELEVENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2018, :108-116
[10]  
Cheng C., 2013, IJCAI, P2605, DOI DOI 10.5555/2540128.2540504