Complementary Attention Gated Network for Pedestrian Trajectory Prediction

被引:0
作者
Duan, Jinghai [1 ]
Wang, Le [2 ]
Long, Chengjiang [3 ]
Zhou, Sanping [2 ]
Zheng, Fang [1 ]
Shi, Liushuai [1 ]
Hua, Gang [4 ]
机构
[1] Xi An Jiao Tong Univ, Sch Software Engn, Xian, Peoples R China
[2] Xi An Jiao Tong Univ, Inst Artificial Intelligence & Robot, Xian, Peoples R China
[3] JD Finance Amer Corp, Mountain View, CA USA
[4] Wormpex AI Res, Bellevue, WA USA
来源
THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2022年
基金
中国博士后科学基金; 国家重点研发计划;
关键词
VISUAL RECOGNITION; CROWDS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pedestrian trajectory prediction is crucial in many practical applications due to the diversity of pedestrian movements, such as social interactions and individual motion behaviors. With similar observable trajectories and social environments, different pedestrians may make completely different future decisions. However, most existing methods only focus on the frequent modal of the trajectory and thus are difficult to generalize to the peculiar scenario, which leads to the decline of the multimodal fitting ability when facing similar scenarios. In this paper, we propose a complementary attention gated network (CAGN) for pedestrian trajectory prediction, in which a dual-path architecture including normal and inverse attention is proposed to capture both frequent and peculiar modals in spatial and temporal patterns, respectively. Specifically, a complementary block is proposed to guide normal and inverse attention, which are then be summed with learnable weights to get attention features by a gated network. Finally, multiple trajectory distributions are estimated based on the fused spatio-temporal attention features due to the multimodality of future trajectory. Experimental results on benchmark datasets, i.e., the ETH, and the UCY, demonstrate that our method outperforms state-of-the-art methods by 13.8% in Average Displacement Error (ADE) and 10.4% in Final Displacement Error (FDE).
引用
收藏
页码:542 / 550
页数:9
相关论文
共 48 条
[31]   Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction [J].
Mohamed, Abduallah ;
Qian, Kun ;
Elhoseiny, Mohamed ;
Claudel, Christian .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :14412-14420
[32]   Trajectory Prediction with Latent Belief Energy-Based Model [J].
Pang, Bo ;
Zhao, Tianyang ;
Xie, Xu ;
Wu, Ying Nian .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :11809-11819
[33]   You'll Never Walk Alone: Modeling Social Behavior for Multi-target Tracking [J].
Pellegrini, S. ;
Ess, A. ;
Schindler, K. ;
van Gool, L. .
2009 IEEE 12TH INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2009, :261-268
[34]   Speaker verification using adapted Gaussian mixture models [J].
Reynolds, DA ;
Quatieri, TF ;
Dunn, RB .
DIGITAL SIGNAL PROCESSING, 2000, 10 (1-3) :19-41
[35]  
Shazeer N., 2017, CORR
[36]   SGCN:Sparse Graph Convolution Network for Pedestrian Trajectory Prediction [J].
Shi, Liushuai ;
Wang, Le ;
Long, Chengjiang ;
Zhou, Sanping ;
Zhou, Mo ;
Niu, Zhenxing ;
Hua, Gang .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :8990-8999
[37]  
Si V. A., 2019, AVSS, P1
[38]   Recursive Social Behavior Graph for Trajectory Prediction [J].
Sun, Jianhua ;
Jiang, Qinhong ;
Lu, Cewu .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :657-666
[39]   Dynamic and Static Context-Aware LSTM for Multi-agent Motion Prediction [J].
Tao, Chaofan ;
Jiang, Qinhong ;
Duan, Lixin ;
Luo, Ping .
COMPUTER VISION - ECCV 2020, PT XXI, 2020, 12366 :547-563
[40]  
Vaswani A, 2017, ADV NEUR IN, V30