Lighter and Better: Low-Rank Decomposed Self-Attention Networks for Next-Item Recommendation

被引:83
作者
Fan, Xinyan [1 ,2 ]
Liu, Zheng [3 ]
Lian, Jianxun [3 ]
Zhao, Wayne Xin [1 ,2 ]
Xie, Xing [3 ]
Wen, Ji-Rong [1 ,2 ]
机构
[1] Renmin Univ China, Gaoling Sch Artificial Intelligence, Beijing, Peoples R China
[2] Renmin Univ China, Beijing Key Lab Big Data Management & Anal Method, Beijing, Peoples R China
[3] Microsoft Res Asia, Beijing, Peoples R China
来源
SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL | 2021年
关键词
Low-Rank Self-Attention; Next-Item Recommendation;
D O I
10.1145/3404835.3462978
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Self-attention networks (SANs) have been intensively applied for sequential recommenders, but they are limited due to: (1) the quadratic complexity and vulnerability to over-parameterization in self-attention; (2) inaccurate modeling of sequential relations between items due to the implicit position encoding. In this work, we propose the low-rank decomposed self-attention networks (LightSANs) to overcome these problems. Particularly, we introduce the low-rank decomposed self-attention, which projects user's historical items into a small constant number of latent interests and leverages item-to-interest interaction to generate the context-aware representation. It scales linearly w.r.t. the user's historical sequence length in terms of time and space, and is more resilient to over-parameterization. Besides, we design the decoupled position encoding, which models the sequential relations between items more precisely. Extensive experimental studies are carried out on three real-world datasets, where LightSANs outperform the existing SANs-based recommenders in terms of both effectiveness and efficiency.
引用
收藏
页码:1733 / 1737
页数:5
相关论文
共 18 条
[1]  
Beltagy I., 2020, Longformer: The Long-Document Transformer, V2004, P05150, DOI DOI 10.48550/ARXIV.2004.05150
[2]  
Choromanski Krzysztof, 2020, P ICLR
[3]  
Dean Jeffrey, 2012, P 26 INT C NEURAL IN, P1223
[4]  
Hidasi Balazs, 2016, P ICLR
[5]   Self-Attentive Sequential Recommendation [J].
Kang, Wang-Cheng ;
McAuley, Julian .
2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, :197-206
[6]  
Katharopoulos A, 2020, PR MACH LEARN RES, V119
[7]  
Ke Guolin, 2020, P INT C LEARN REPR
[8]   Neural Attentive Session-based Recommendation [J].
Li, Jing ;
Ren, Pengjie ;
Chen, Zhumin ;
Ren, Zhaochun ;
Lian, Tao ;
Ma, Jun .
CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, :1419-1428
[9]   Parallel Implementation of Chaos Neural Networks for an Embedded GPU [J].
Liu, Zhongda ;
Murakami, Takeshi ;
Kawamura, Satoshi ;
Yoshida, Hitoaki .
2019 IEEE 10TH INTERNATIONAL CONFERENCE ON AWARENESS SCIENCE AND TECHNOLOGY (ICAST 2019), 2019, :34-39
[10]  
Rendle Steffen, 2010, P 19 INT C WORLD WID, P811, DOI DOI 10.1145/1772690.1772773