Path Planning of Unmanned Helicopter in Complex Dynamic Environment Based on State-Coded Deep Q-Network

被引:4
|
作者
Yao, Jiangyi [1 ]
Li, Xiongwei [1 ]
Zhang, Yang [1 ]
Ji, Jingyu [2 ]
Wang, Yanchao [1 ]
Liu, Yicen [3 ]
机构
[1] Army Engn Univ, Equipment Simulat Training Ctr, Shijiazhuang Campus, Shijiazhuang 050003, Hebei, Peoples R China
[2] Army Engn Univ, Dept UAV, Shijiazhuang Campus, Shijiazhuang 050003, Hebei, Peoples R China
[3] State Key Lab Blind Signal Proc, Chengdu 610000, Peoples R China
来源
SYMMETRY-BASEL | 2022年 / 14卷 / 05期
基金
中国国家自然科学基金;
关键词
path planning; deep reinforcement learning; dynamic reward function; status code; LARGE-SCALE; A-ASTERISK; ALGORITHM; VEHICLE;
D O I
10.3390/sym14050856
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Unmanned helicopters (UH) can avoid radar detection by flying at ultra-low altitudes; thus, they have been widely used in the battlefield. The flight safety of UH is seriously affected by moving obstacles such as flocks of birds in low airspace. Therefore, an algorithm that can plan a safe path to UH is urgently needed. Due to the strong randomness of the movement of bird flocks, the existing path planning algorithms are incompetent for this task. To solve this problem, a state-coded deep Q-network (SC-DQN) algorithm with symmetric properties is proposed, which can effectively avoid randomly moving obstacles and plan a safe path for UH. First, a dynamic reward function is designed to give UH appropriate rewards in real time, so as to improve the sparse reward problem. Then, a state-coding scheme is proposed, which uses binary Boolean expression to encode the environment state to compress environment state space. The encoded state is used as the input to the deep learning network, which is an important improvement to the traditional algorithm. Experimental results show that the SC-DQN algorithm can help UH avoid the moving obstacles to unknown motion status in the environment safely and effectively and successfully complete the raid task.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Path Planning of Unmanned Helicopter in Complex Environment Based on Heuristic Deep Q-Network
    Yao, Jiangyi
    Li, Xiongwei
    Zhang, Yang
    Ji, Jingyu
    Wang, Yanchao
    Liu, Yicen
    INTERNATIONAL JOURNAL OF AEROSPACE ENGINEERING, 2022, 2022
  • [2] Path planning for unmanned vehicle reconnaissance based on deep Q-network
    Xia, Yuqi
    Huang, Yanyan
    Chen, Qia
    Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics, 2024, 46 (09): : 3070 - 3081
  • [3] A Stealth-Distance Dynamic Weight Deep Q-Network Algorithm for Three-Dimensional Path Planning of Unmanned Aerial Helicopter
    Wang, Zeyang
    Huang, Jun
    Yi, Mingxu
    AEROSPACE, 2023, 10 (08)
  • [4] Multirobot Coverage Path Planning Based on Deep Q-Network in Unknown Environment
    Li, Wenhao
    Zhao, Tao
    Dian, Songyi
    JOURNAL OF ROBOTICS, 2022, 2022
  • [5] AGV Path Planning with Dynamic Obstacles Based on Deep Q-Network and Distributed Training
    Xie, Tingbo
    Yao, Xifan
    Jiang, Zhenhong
    Meng, Junting
    INTERNATIONAL JOURNAL OF PRECISION ENGINEERING AND MANUFACTURING-GREEN TECHNOLOGY, 2025,
  • [6] Dynamic Path Planning Scheme for OHT in AMHS Based on Map Information Double Deep Q-Network
    Ao, Qi
    Zhou, Yue
    Guo, Wei
    Wang, Wenguang
    Ye, Ying
    ELECTRONICS, 2024, 13 (22)
  • [7] Multiple Unmanned Aerial Vehicle Autonomous Path Planning Algorithm Based on Whale-Inspired Deep Q-Network
    Wang, Wenshan
    Zhang, Guoyin
    Da, Qingan
    Lu, Dan
    Zhao, Yingnan
    Li, Sizhao
    Lang, Dapeng
    DRONES, 2023, 7 (09)
  • [8] Path planning of mobile robot in unknown dynamic continuous environment using reward-modified deep Q-network
    Huang, Runnan
    Qin, Chengxuan
    Li, Jian Ling
    Lan, Xuejing
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2023, 44 (03): : 1570 - 1587
  • [9] Path planning of mobile robot based on improved double deep Q-network algorithm
    Wang, Zhenggang
    Song, Shuhong
    Cheng, Shenghui
    FRONTIERS IN NEUROROBOTICS, 2025, 19
  • [10] Transport robot path planning based on an advantage dueling double deep Q-network
    He Q.
    Wang Q.
    Li J.
    Wang Z.
    Wang T.
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2022, 62 (11): : 1751 - 1757