Prioritized Environment Configuration for Drone Control with Deep Reinforcement Learning

被引:3
|
作者
Jang, Sooyoung [1 ]
Choi, Changbeom [2 ]
机构
[1] Elect & Telecommun Res Inst ETRI, Intelligence Convergence Res Lab, Daejeon, South Korea
[2] Hanbat Natl Univ, Dept Comp Engn, Daejeon, South Korea
关键词
Deep Reinforcement Learning; Machine Learning; Prioritized Environment Configuration; Environment; Initialization; Drone Control;
D O I
10.22967/HCIS.2022.12.002
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In reinforcement learning, first, the agent collects experiences by interacting with the environment through trial-and-errors (experience collection stage) and then learns from the collected experiences (learning stage). This two-stage training process repeats until the agent solves a given task and requires a lot of experience, computation power, and time for training the agent. Therefore, many studies are conducted to improve the training speed and performance to mitigate them, focusing on the learning stage. This paper focuses on the experience collection stage and proposes a prioritized environment configuration that prioritizes and stochastically samples the effective configuration for initializing the environment for every episode. Therefore, we can provide the environments initialized with the configuration suitable for effective experience collection to the agent. The proposed algorithm can complement the reinforcement learning algorithms that focus on the learning stage. We have shown speed and performance improvement by applying the prioritized environment configuration to an autonomous drone flight simulator. In addition, the results show that the proposed algorithm works well with both on-policy and off-policy reinforcement learning algorithms in distributed framework with multiple workers.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Drone Deep Reinforcement Learning: A Review
    Azar, Ahmad Taher
    Koubaa, Anis
    Ali Mohamed, Nada
    Ibrahim, Habiba A.
    Ibrahim, Zahra Fathy
    Kazim, Muhammad
    Ammar, Adel
    Benjdira, Bilel
    Khamis, Alaa M.
    Hameed, Ibrahim A.
    Casalino, Gabriella
    ELECTRONICS, 2021, 10 (09)
  • [2] Deep Reinforcement Learning for Drone Delivery
    Munoz, Guillem
    Barrado, Cristina
    Cetin, Ender
    Salami, Esther
    DRONES, 2019, 3 (03) : 1 - 19
  • [3] Limit Action Space to Enhance Drone Control with Deep Reinforcement Learning
    Jang, Sooyoung
    Park, Noh-Sam
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1212 - 1215
  • [4] Drone Altitude Control with Reinforcement Learning
    Fu, Xilin
    Tay, Eng Hock Francis
    Hu, Junru
    Zhang, Yingnan
    Ding, Yi
    PROCEEDINGS OF 2022 INTERNATIONAL CONFERENCE ON AUTONOMOUS UNMANNED SYSTEMS, ICAUS 2022, 2023, 1010 : 590 - 594
  • [5] The Use of Deep Reinforcement Learning for Flying a Drone
    Domitran, Sandro
    Babac, Marina Bagic
    JOURNAL OF INFORMATION SCIENCE AND ENGINEERING, 2021, 37 (05) : 1165 - 1176
  • [6] Autonomous Drone Racing with Deep Reinforcement Learning
    Song, Yunlong
    Steinweg, Mats
    Kaufmann, Elia
    Scaramuzza, Davide
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 1205 - 1212
  • [7] Autonomous drone interception with Deep Reinforcement Learning
    Bertoin, David
    Gauffriau, Adrien
    Grasset, Damien
    Gupta, Jayant Sen
    CEUR Workshop Proceedings, 2022, 3173
  • [8] Collision avoidance for a small drone with a monocular camera using deep reinforcement learning in an indoor environment
    Kim M.
    Kim J.
    Jung M.
    Oh H.
    Journal of Institute of Control, Robotics and Systems, 2020, 26 (06) : 399 - 411
  • [9] Continuous drone control using deep reinforcement learning for frontal view person shooting
    Nikolaos Passalis
    Anastasios Tefas
    Neural Computing and Applications, 2020, 32 : 4227 - 4238
  • [10] Continuous drone control using deep reinforcement learning for frontal view person shooting
    Passalis, Nikolaos
    Tefas, Anastasios
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (09): : 4227 - 4238