Social NSTransformers: Low-Quality Pedestrian Trajectory Prediction

被引:4
作者
Jiang Z. [1 ]
Ma Y. [1 ]
Shi B. [2 ]
Lu X. [1 ]
Xing J. [2 ]
Goncalves N. [3 ]
Jin B. [3 ]
机构
[1] Xi'an Jiaotong-Liverpool University, School of Advanced Technology (SAT), Suzhou
[2] Northeast Forestry University, College of Computer and Control Engineering, Harbin
[3] University of Coimbra, Institute of Systems and Robotics, Department of Electrical and Computer Engineering, Coimbra
来源
IEEE Transactions on Artificial Intelligence | 2024年 / 5卷 / 11期
关键词
Enhanced loss function; generative adversarial network (GAN); nonstationary transformers (NSTransformers); pedestrian trajectory prediction;
D O I
10.1109/TAI.2024.3421175
中图分类号
学科分类号
摘要
This article introduces a novel model for low-quality pedestrian trajectory prediction, the social nonstationary transformers (NSTransformers), that merges the strengths of NSTransformers and spatiotemporal graph transformer (STAR). The model can capture social interaction cues among pedestrians and integrate features across spatial and temporal dimensions to enhance the precision and resilience of trajectory predictions. We also propose an enhanced loss function that combines diversity loss with logarithmic root mean squared error (log-RMSE) to guarantee the reasonableness and diversity of the generated trajectories. This design adapts well to complex pedestrian interaction scenarios, thereby improving the reliability and accuracy of trajectory prediction. Furthermore, we integrate a generative adversarial network (GAN) to model the randomness inherent in pedestrian trajectories. Compared to the conventional standard Gaussian distribution, our GAN approach better simulates the intricate distribution found in pedestrian trajectories, enhancing the trajectory prediction's diversity and robustness. Experimental results reveal that our model outperforms several state-of-the-art methods. This research opens the avenue for future exploration in low-quality pedestrian trajectory prediction. © 2024 IEEE.
引用
收藏
页码:5575 / 5588
页数:13
相关论文
共 51 条
[1]  
Jiang Z., Et al., Noise interference reduction in vision module of intelligent plant cultivation robot using better cycle GAN, IEEE Sensors J., 22, 11, pp. 11045-11055, (2022)
[2]  
Song Y., He Z., Qian H., Du X., Vision transformers for single image dehazing, IEEE Trans. Image Process., 32, pp. 1927-1941, (2023)
[3]  
Mehta N., Murala S., Image super-resolution with content-Aware feature processing, IEEE Trans. Artif. Intell., 5, 1, pp. 179-191, (2024)
[4]  
Barman T., Deka B., A deep learning-based joint image superresolution and deblurring framework, IEEE Trans. Artif. Intell., 5, 6, pp. 3160-3173, (2024)
[5]  
Wang H., Peng L., Sun Y., Wan Z., Wang Y., Cao Y., Brightness perceiving for recursive low-light image enhancement, IEEE Trans. Artif. Intell., 5, 6, pp. 3034-3045, (2024)
[6]  
Sharma T., Verma N.K., Adaptive interval type-2 fuzzy filter: An AI agent for handling uncertainties to preserve image naturalness, IEEE Trans. Artif. Intell., 2, 1, pp. 83-92, (2021)
[7]  
Nath K., Bera M.K., Jagannathan S., Concurrent learning-based neuroadaptive robust tracking control of wheeled mobile robot: An event-Triggered design, IEEE Trans. Artif. Intell., 4, 6, pp. 1514-1525, (2023)
[8]  
He X., Lv C., Robotic control in adversarial and sparse reward environments: A robust goal-conditioned reinforcement learning approach, IEEE Trans. Artif. Intell., 5, 1, pp. 244-253, (2024)
[9]  
Li X., Li X., Li Z., Xiong X., Khyam M.O., Sun C., Robust vehicle detection in high-resolution aerial images with imbalanced data, IEEE Trans. Artif. Intell., 2, 3, pp. 238-250, (2021)
[10]  
Zhang Q., Li C., Semantic SLAM for mobile robots in dynamic environments based on visual camera sensors, Meas. Sci. Technol., 34, 8, (2023)