Sparse Attentive Memory Network for Click-through Rate Prediction with Long Sequences

被引:4
作者
Lin, Qianying [1 ]
Zhou, Wen-Ji [1 ]
Wang, Yanshi [1 ]
Da, Qing [1 ]
Chen, Qing-Guo [1 ]
Wang, Bing [1 ]
机构
[1] Alibaba Grp, Hangzhou, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022 | 2022年
关键词
Sequential Recommenders; Long User Behavior Modeling; Long Sequences; Click-through Rate Prediction; Memory Networks;
D O I
10.1145/3511808.3557095
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sequential recommendation predicts users' next behaviors with their historical interactions. Recommending with longer sequences improves recommendation accuracy and increases the degree of personalization. As sequences get longer, existing works have not yet addressed the following two main challenges. Firstly, modeling long-range intra-sequence dependency is difficult with increasing sequence lengths. Secondly, it requires efficient memory and computational speeds. In this paper, we propose a Sparse Attentive Memory (SAM) network for long sequential user behavior modeling. SAM supports efficient training and real-time inference for user behavior sequences with lengths on the scale of thousands. In SAM, we model the target item as the query and the long sequence as the knowledge database, where the former continuously elicits relevant information from the latter. SAM simultaneously models targetsequence dependencies and long-range intra-sequence dependencies with O(L) complexity and O(1) number of sequential updates, which can only be achieved by the self-attention mechanism with O(L-2) complexity. Extensive empirical results demonstrate that our proposed solution is effective not only in long user behavior modeling but also on short sequences modeling. Implemented on sequences of length 1000, SAM is successfully deployed on one of the largest international E-commerce platforms. This inference time is within 30ms, with a substantial 7.30% click-through rate improvement for the online A/B test. To the best of our knowledge, it is the first end-to-end long user sequence modeling framework that models intra-sequence and target-sequence dependencies with the aforementioned degree of efficiency and successfully deployed on a large-scale real-time industrial recommender system.
引用
收藏
页码:3312 / 3321
页数:10
相关论文
共 51 条
[1]  
[Anonymous], 2019, P 1 INT WORKSH DEEP
[2]  
[Anonymous], 2017, ARXIV171003348
[3]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[4]  
Cecilia Jose M., 2009, PARCO
[5]  
Chan W., 2015, Listen, attend and spell
[6]   An Attentive Survey of Attention Models [J].
Chaudhari, Sneha ;
Mithal, Varun ;
Polatkan, Gungor ;
Ramanath, Rohan .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (05)
[7]   Sequential Recommendation with User Memory Networks [J].
Chen, Xu ;
Xu, Hongteng ;
Zhang, Yongfeng ;
Tang, Jiaxi ;
Cao, Yixin ;
Qin, Zheng ;
Zha, Hongyuan .
WSDM'18: PROCEEDINGS OF THE ELEVENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2018, :108-116
[8]   Deep Neural Networks for YouTube Recommendations [J].
Covington, Paul ;
Adams, Jay ;
Sargin, Emre .
PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, :191-198
[9]  
Cui Y., 2016, Attention-over-attention neural networks for reading comprehension
[10]   Optimizing Sparse Matrix Operations on GPUs using Merge Path [J].
Dalton, Steven ;
Olson, Luke ;
Baxter, Sean ;
Merrill, Duane ;
Garland, Michael .
2015 IEEE 29TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2015, :407-416