Highly Interactive Self-Supervised Learning for Multi-Modal Trajectory Prediction

被引:0
|
作者
Xie, Wenda [1 ]
Liu, Yahui [1 ]
Zhao, Hongxia [1 ]
Guo, Chao [1 ]
Dai, Xingyuan [1 ,2 ]
Lv, Yisheng [1 ,3 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai 200232, Peoples R China
[3] Changan Univ, Minist Educ, Engn Res Ctr Highway Infrastruct Digitalizat, Xian 710064, Peoples R China
来源
IFAC PAPERSONLINE | 2024年 / 58卷 / 10期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Automatic driving; Self-supervised learning; Trajectory prediction; Deep learning; Intelligent Transportation;
D O I
10.1016/j.ifacol.2024.07.345
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To ensure the safety of autonomous vehicles, trajectory prediction is critical as it enables vehicles to anticipate the movements of surrounding agents, thereby facilitating the planning of secure and strategic driving routes. However, striking a trade-off between predictive accuracy and training costs has always been an intricate challenge. This paper introduces a groundbreaking framework for trajectory prediction known as Highly Interactive Self-Supervised Learning (HI-SSL), a methodology based on self-supervised learning (SSL) that has yet to be thoroughly investigated in the realm of trajectory prediction. The cornerstone of the aforementioned framework is Interactive Masking, which leverages a novel trajectory masking strategy facilitating self-supervised learning tasks that not only enhance prediction accuracy but also eliminate the need for manual annotations. Experiments conducted on the Argoverse motion forecasting dataset demonstrate that our approach achieves competitive performance to prior methods that depend on supervised learning without additional annotation costs. Copyright (C) 2024 The Authors. This is an open access article under the CC BY-NC-ND license (https://creativecommons.org/licenses/by-nc-nd/4.0/)
引用
收藏
页码:231 / 236
页数:6
相关论文
共 50 条
  • [41] AUDIO-VISUAL SPEECH ENHANCEMENT AND SEPARATION BY UTILIZING MULTI-MODAL SELF-SUPERVISED EMBEDDINGS
    Chern, I-Chun
    Hung, Kuo-Hsuan
    Chen, Yi-Ting
    Hussain, Tassadaq
    Gogate, Mandar
    Hussain, Amir
    Tsao, Yu
    Hou, Jen-Cheng
    2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [42] Towards Multi-modal Self-supervised Video and Ultrasound Pose Estimation for Laparoscopic Liver Surgery
    Montana-Brown, Nina
    Ramalhinho, Joao
    Koo, Bongjin
    Allam, Moustafa
    Davidson, Brian
    Gurusamy, Kurinchi
    Hu, Yipeng
    Clarkson, Matthew J.
    SIMPLIFYING MEDICAL ULTRASOUND, ASMUS 2022, 2022, 13565 : 183 - 192
  • [43] Self-supervised learning for modal transfer of brain imaging
    Cheng, Dapeng
    Chen, Chao
    Yanyan, Mao
    You, Panlu
    Huang, Xingdan
    Gai, Jiale
    Zhao, Feng
    Mao, Ning
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [44] A Lightweight Multi-Modal Vehicle Trajectory Prediction Algorithm
    Li Z.
    Sun H.
    Hao Z.
    Xiao D.
    Hsi-An Chiao Tung Ta Hsueh/Journal of Xi'an Jiaotong University, 2024, 58 (06): : 14 - 23
  • [45] Multi-Task Self-Supervised Learning for Script Event Prediction
    Zhou, Bo
    Chen, Yubo
    Liu, Kang
    Zhao, Jun
    Xu, Jiexin
    Jiang, Xiaojian
    Li, Jinlong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3662 - 3666
  • [46] Interactive Contrastive Learning for Self-Supervised Entity Alignment
    Zeng, Kaisheng
    Dong, Zhenhao
    Hou, Lei
    Cao, Yixin
    Hu, Minghao
    Yu, Jifan
    Lv, Xin
    Cao, Lei
    Wang, Xin
    Liu, Haozhuang
    Huang, Yi
    Feng, Junlan
    Wan, Jing
    Li, Juanzi
    Feng, Ling
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2465 - 2475
  • [47] Comprehensive Semi-Supervised Multi-Modal Learning
    Yang, Yang
    Wang, Ke-Tao
    Zhan, De-Chuan
    Xiong, Hui
    Jiang, Yuan
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4092 - 4098
  • [48] Supervised Multi-modal Dictionary Learning for Clothing Representation
    Zhao, Qilu
    Wang, Jiayan
    Li, Zongmin
    PROCEEDINGS OF THE FIFTEENTH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS - MVA2017, 2017, : 51 - 54
  • [49] Multi-Modal Sentiment Classification With Independent and Interactive Knowledge via Semi-Supervised Learning
    Zhang, Dong
    Li, Shoushan
    Zhu, Qiaoming
    Zhou, Guodong
    IEEE ACCESS, 2020, 8 : 22945 - 22954
  • [50] Self-supervised Learning for Unintentional Action Prediction
    Zatsarynna, Olga
    Abu Farha, Yazan
    Gall, Juergen
    PATTERN RECOGNITION, DAGM GCPR 2022, 2022, 13485 : 429 - 444