Monocular Camera-Based Complex Obstacle Avoidance via Efficient Deep Reinforcement Learning

被引:19
作者
Ding, Jianchuan [1 ,2 ]
Gao, Lingping [1 ,3 ]
Liu, Wenxi [4 ]
Piao, Haiyin [5 ]
Pan, Jia [6 ]
Du, Zhenjun [7 ]
Yang, Xin [8 ]
Yin, Baocai [8 ]
机构
[1] Dalian Univ Technol, Sch Comp Sci, Dalian 116024, Peoples R China
[2] Hebei Univ Water Resources & Elect Engn, Sch Comp Sci & Informat Engn, Cangzhou 061016, Peoples R China
[3] Alibaba Grp, Hangzhou 310000, Peoples R China
[4] Fuzhou Univ, Coll Math & Comp Sci, Fuzhou 350108, Peoples R China
[5] Northwestern Polytech Univ, Sch Elect & Informat, Xian 710072, Peoples R China
[6] Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
[7] SIASUN Robot & Automat Co Ltd, Shenyang 110168, Peoples R China
[8] Dalian Univ Technol, Dalian 116024, Peoples R China
基金
中国国家自然科学基金;
关键词
Collision avoidance; Robots; Robot sensing systems; Semantics; Measurement by laser beam; Cameras; Sensors; Deep reinforcement learning; obstacle avoidance; robot vision; robot navigation; NAVIGATION;
D O I
10.1109/TCSVT.2022.3203974
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep reinforcement learning has achieved great success in laser-based collision avoidance works because the laser can sense accurate depth information without too much redundant data, which can maintain the robustness of the algorithm when it is migrated from the simulation environment to the real world. However, high-cost laser devices are not only difficult to deploy for a large scale of robots but also demonstrate unsatisfactory robustness towards the complex obstacles, including irregular obstacles, e.g., tables, chairs, and shelves, as well as complex ground and special materials. In this paper, we propose a novel monocular camera-based complex obstacle avoidance framework. Particularly, we innovatively transform the captured RGB images to pseudo-laser measurements for efficient deep reinforcement learning. Compared to the traditional laser measurement captured at a certain height that only contains one-dimensional distance information away from the neighboring obstacles, our proposed pseudo-laser measurement fuses the depth and semantic information of the captured RGB image, which makes our method effective for complex obstacles. We also design a feature extraction guidance module to weight the input pseudo-laser measurement, and the agent has more reasonable attention for the current state, which is conducive to improving the accuracy and efficiency of the obstacle avoidance policy. Besides, we adaptively add the synthesized noise to the laser measurement during the training stage to decrease the sim-to-real gap and increase the robustness of our model in the real environment. Finally, the experimental results show that our framework achieves state-of-the-art performance in several virtual and real-world scenarios.
引用
收藏
页码:756 / 770
页数:15
相关论文
共 58 条
[11]   The dynamic window approach to collision avoidance [J].
Fox, D ;
Burgard, W ;
Thrun, S .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 1997, 4 (01) :23-33
[12]  
Fragkiadaki K, 2016, Arxiv, DOI arXiv:1511.07404
[13]  
Howard AG, 2017, Arxiv, DOI arXiv:1704.04861
[14]   A Vision-based Irregular Obstacle Avoidance Framework via Deep Reinforcement Learning [J].
Gao, Lingping ;
Ding, Jianchuan ;
Liu, Wenxi ;
Piao, Haiyin ;
Wang, Yuxin ;
Yang, Xin ;
Yin, Baocai .
2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, :9262-9269
[15]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[16]   A Machine Learning Approach to Visual Perception of Forest Trails for Mobile Robots [J].
Giusti, Alessandro ;
Guzzi, Jerome ;
Ciresan, Dan C. ;
He, Fang-Lin ;
Rodriguez, Juan P. ;
Fontana, Flavio ;
Faessler, Matthias ;
Forster, Christian ;
Schmidhuber, Jurgen ;
Di Caro, Gianni ;
Scaramuzza, Davide ;
Gambardella, Luca M. .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2016, 1 (02) :661-667
[17]   SplitNet: Sim2Sim and Task2Task Transfer for Embodied Visual Navigation [J].
Gordon, Daniel ;
Kadian, Abhishek ;
Parikh, Devi ;
Hoffman, Judy ;
Batra, Dhruv .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1022-1031
[18]   Deep Visual MPC-Policy Learning for Navigation [J].
Hirose, Noriaki ;
Xia, Fei ;
Martin-Martin, Roberto ;
Sadeghian, Amir ;
Savarese, Silvio .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04) :3184-3191
[19]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[20]   Global elimination algorithm and architecture design for fast block matching motion estimation [J].
Huang, YW ;
Chien, SY ;
Hsieh, BY ;
Chen, LG .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2004, 14 (06) :898-907