Spatial-Temporal Cross-View Contrastive Pre-Training for Check-in Sequence Representation Learning

被引:0
作者
Gong, Letian [1 ,2 ]
Wan, Huaiyu [1 ,2 ]
Guo, Shengnan [1 ,2 ]
Li, Xiucheng [3 ]
Lin, Yan [1 ,2 ]
Zheng, Erwen [1 ,2 ]
Wang, Tianyi [1 ,2 ]
Zhou, Zeyu [1 ,2 ]
Lin, Youfang [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Key Lab Big Data & Artificial Intelligence Transpo, Minist Educ, Beijing 100044, Peoples R China
[2] CAAC, Key Lab Intelligent Passenger Serv Civil Aviat, Beijing 101318, Peoples R China
[3] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Semantics; Trajectory; Predictive models; Uncertainty; Task analysis; Noise; Data mining; Check-in sequence; contrastive cluster; representation learning; spatial-temporal cross-view;
D O I
10.1109/TKDE.2024.3434565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rapid growth of location-based services (LBS) has yielded massive amounts of data on human mobility. Effectively extracting meaningful representations for user-generated check-in sequences is pivotal for facilitating various downstream services. However, the user-generated check-in data are simultaneously influenced by the surrounding objective circumstances and the user's subjective intention. Specifically, the temporal uncertainty and spatial diversity exhibited in check-in data make it difficult to capture the macroscopic spatial-temporal patterns of users and to understand the semantics of user mobility activities. Furthermore, the distinct characteristics of the temporal and spatial information in check-in sequences call for an effective fusion method to incorporate these two types of information. In this paper, we propose a novel Spatial-Temporal Cross-view Contrastive Representation (STCCR) framework for check-in sequence representation learning. Specifically, STCCR addresses the above challenges by employing self-supervision from "spatial topic" and "temporal intention" views, facilitating effective fusion of spatial and temporal information at the semantic level. Besides, STCCR leverages contrastive clustering to uncover users' shared spatial topics from diverse mobility activities, while employing angular momentum contrast to mitigate the impact of temporal uncertainty and noise. We extensively evaluate STCCR on three real-world datasets and demonstrate its superior performance across three downstream tasks.
引用
收藏
页码:9308 / 9321
页数:14
相关论文
共 60 条
[41]   Heterogeneous Graph Attention Network [J].
Wang, Xiao ;
Ji, Houye ;
Shi, Chuan ;
Wang, Bai ;
Cui, Peng ;
Yu, P. ;
Ye, Yanfang .
WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, :2022-2032
[42]  
Wang Y., 2020, arXiv
[43]   Exploring graph capsual network for graph classification [J].
Wang, Ying ;
Wang, Hongji ;
Jin, Hui ;
Huang, Xinrui ;
Wang, Xin .
INFORMATION SCIENCES, 2021, 581 :932-950
[44]   Advanced graph and sequence neural networks for molecular property prediction and drug discovery [J].
Wang, Zhengyang ;
Liu, Meng ;
Luo, Youzhi ;
Xu, Zhao ;
Xie, Yaochen ;
Wang, Limei ;
Cai, Lei ;
Qi, Qi ;
Yuan, Zhuoning ;
Yang, Tianbao ;
Ji, Shuiwang .
BIOINFORMATICS, 2022, 38 (09) :2579-2586
[45]  
Wasserman S, 1994, Social network analysis: Methods and applications, DOI 10.1017/CBO9780511815478
[46]  
Wei T., 2022, arXiv
[47]  
Xu K, 2018, arXiv
[48]   Model-Agnostic Augmentation for Accurate Graph Classification [J].
Yoo, Jaemin ;
Shim, Sooyeon ;
Kang, U. .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :1281-1291
[49]  
You Y, 2020, ADV NEUR IN, V33
[50]  
Zaremba W, 2015, Arxiv, DOI arXiv:1409.2329