Robust User Behavioral Sequence Representation via Multi-scale Stochastic Distribution Prediction

被引:4
作者
Fu, Chilin [1 ]
Wu, Weichang [1 ]
Zhang, Xiaolu [1 ]
Hu, Jun [1 ]
Wang, Jing [1 ]
Zhou, Jun [1 ]
机构
[1] Ant Grp, Hangzhou, Zhejiang, Peoples R China
来源
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023 | 2023年
关键词
machine learning; sequential data mining; representation learning;
D O I
10.1145/3583780.3614714
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
User behavior representation learned by self-supervised pre-training tasks is widely used in various domains and applications. Conventional methods usually follow the methodology in Natural Language Processing (NLP) to set the pre-training tasks. They either randomly mask some of the behaviors in the sequence and predict the masked ones or predict the next k. behaviors. These methods fit for text sequence, in which the tokens are sequentially arranged subject to linguistic criterion. However, the user behavior sequences can be stochastic with noise and randomness. The same paradigm is intractable for learning a robust user behavioral representation. Though the next user behavior can be stochastic, the behavior distribution over a period of time is much more stable and less noisy. Based on this, we propose a Multi-scale Stochastic Distribution Prediction (MSDP) algorithm for learning robust user behavioral sequence representation. Instead of using predictions on concrete behavior as pre-training tasks, we take the prediction on user's behaviors distribution over a period of time as the self-supervision signal. Moreover, inspired by the recent success of the multi-task prompt training method on Large Language Models (LLM), we propose using the window size of the predicted time period as a prompt, enabling the model to learn user behavior representations that can be applied to prediction tasks across various future time periods. We generate different window size prompts through stochastic sampling. It effectively improves the generalization capability of the learned sequence representation. Extensive experiments demonstrate that our approach can learn robust user behavior representation successfully, which significantly outperforms state-of-the-art (SOTA) baselines.
引用
收藏
页码:4567 / 4573
页数:7
相关论文
共 18 条
[1]  
Adiwardana D., 2020, arXiv preprint arXiv:2001.09977
[2]  
Bachman P, 2019, ADV NEUR IN, V32
[3]   Exploring Simple Siamese Representation Learning [J].
Chen, Xinlei ;
He, Kaiming .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :15745-15753
[4]   Learning Transferable User Representations with Sequential Behaviors via Contrastive Pre-training [J].
Cheng, Mingyue ;
Yuan, Fajie ;
Liu, Qi ;
Xin, Xin ;
Chen, Enhong .
2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, :51-60
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]  
Dongwook Lee, 2019, arXiv
[7]  
Gu J, 2021, AAAI CONF ARTIF INTE, V35, P4063
[8]   PinnerFormer: Sequence Modeling for User Representation at Pinterest [J].
Pancha, Nikil ;
Zhai, Andrew ;
Leskovec, Jure ;
Rosenberg, Charles .
PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, :3702-3712
[9]  
Raffel Colin, 2019, J MACH LEARN RES
[10]  
Shin Kyuyong, 2021, ARXIV210600573