Turning Dross Into Gold Loss: is BERT4Rec really better than SASRec?

被引:21
作者
Klenitskiy, Anton [1 ]
Vasilev, Alexey [1 ]
机构
[1] Sber, AI Lab, Moscow, Russia
来源
PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023 | 2023年
关键词
recommender systems; sequential recsys; BERT4Rec; SASRec;
D O I
10.1145/3604915.3610644
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently sequential recommendations and next-item prediction task has become increasingly popular in the field of recommender systems. Currently, two state-of-the-art baselines are Transformer-based models SASRec and BERT4Rec. Over the past few years, there have been quite a few publications comparing these two algorithms and proposing new state-of-the-art models. In most of the publications, BERT4Rec achieves better performance than SASRec. But BERT4Rec uses cross-entropy over softmax for all items, while SASRec uses negative sampling and calculates binary cross-entropy loss for one positive and one negative item. In our work, we show that if both models are trained with the same loss, which is used by BERT4Rec, then SASRec will significantly outperform BERT4Rec both in terms of quality and training speed. In addition, we show that SASRec could be effectively trained with negative sampling and still outperform BERT4Rec, but the number of negative examples should be much larger than one.
引用
收藏
页码:1120 / 1125
页数:6
相关论文
共 35 条
[1]  
Akiba T, 2019, Arxiv, DOI [arXiv:1907.10902, DOI 10.48550/ARXIV.1907.10902]
[2]  
[Anonymous], 2018, Fast python collaborative filtering for implicit datasets
[3]  
antklen, antklen/sasrec-bert4rec-recsys23
[4]  
Asghar N, 2016, Arxiv, DOI arXiv:1605.05362
[5]   On Target Item Sampling in Offline Recommender System Evaluation [J].
Canamares, Rocio ;
Castells, Pablo .
RECSYS 2020: 14TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2020, :259-268
[6]   Denoising Self-Attentive Sequential Recommendation [J].
Chen, Huiyuan ;
Lin, Yusan ;
Pan, Menghai ;
Wang, Lan ;
Yeh, Chin-Chia Michael ;
Li, Xiaoting ;
Zheng, Yan ;
Wang, Fei ;
Yang, Hao .
PROCEEDINGS OF THE 16TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2022, 2022, :92-101
[7]   A Case Study on Sampling Strategies for Evaluating Neural Sequential Item Recommendation Models [J].
Dallmann, Alexander ;
Zoller, Daniel ;
Hotho, Andreas .
15TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS 2021), 2021, :505-514
[8]   Contrastive Learning with Bidirectional Transformers for Sequential Recommendation [J].
Du, Hanwen ;
Shi, Hui ;
Zhao, Pengpeng ;
Wang, Deqing ;
Sheng, Victor S. ;
Liu, Yanchi ;
Liu, Guanfeng ;
Zhao, Lei .
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, :396-405
[9]   Lighter and Better: Low-Rank Decomposed Self-Attention Networks for Next-Item Recommendation [J].
Fan, Xinyan ;
Liu, Zheng ;
Lian, Jianxun ;
Zhao, Wayne Xin ;
Xie, Xing ;
Wen, Ji-Rong .
SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, :1733-1737
[10]  
FeiSun, FeiSun/BERT4Rec