A Deep Deterministic Policy Gradient Approach for Vehicle Speed Tracking Control With a Robotic Driver

被引:23
|
作者
Hao, Gaofeng [1 ]
Fu, Zhuang [1 ]
Feng, Xin [1 ]
Gong, Zening [1 ]
Chen, Peng [2 ]
Wang, Dan [2 ]
Wang, Weibin [2 ]
Si, Yang [2 ]
机构
[1] Shanghai Jiao Tong Univ, State Key Lab Mech Syst & Vibrat, Shanghai 200240, Peoples R China
[2] Pan Asia Tech Automot Ctr PATAC, Shanghai 201201, Peoples R China
基金
中国国家自然科学基金;
关键词
Robots; Testing; Dynamometers; Training; Automobiles; Aerospace electronics; Resistance; Deep deterministic policy gradient (DDPG); network exploration; reinforcement learning (RL); replay buffer; robotic driver; vehicle speed tracking control;
D O I
10.1109/TASE.2021.3088004
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In performance tests, replacing humans with robotic drivers has many advantages, such as high efficiency and high security. To realize the vehicle speed tracking control with a robotic driver, this article proposes a novel deep reinforcement learning (DRL) approach based on deep deterministic policy gradient (DDPG). Specifically, the design of the approach includes state space, action space, reward function, and control algorithm. Then, to shorten the training time, the proposed approach utilizes the basic fundamental relationship between vehicle speed and pedal opening to intervene in network exploration. Furthermore, to solve speed fluctuations in low-speed sections, the replay buffer is optimized by adding weighted training samples. Experiments are conducted on fifteen cars, and results show that the proposed algorithm can effectively control the vehicle speed. Generally, it only needs three or four episodes of training to meet the requirements. Compared with the Segment-PID method, the proposed method has a smoother speed and fewer overbound times.
引用
收藏
页码:2514 / 2525
页数:12
相关论文
共 43 条
  • [1] Deep Deterministic Policy Gradient for High-Speed Train Trajectory Optimization
    Ning, Lingbin
    Zhou, Min
    Hou, Zhuopu
    Goverde, Rob M. P.
    Wang, Fei-Yue
    Dong, Hairong
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 11562 - 11574
  • [2] Deep Deterministic Policy Gradient With Prioritized Sampling for Power Control
    Zhou, Shiyang
    Cheng, Yufan
    Lei, Xia
    Duan, Huanhuan
    IEEE ACCESS, 2020, 8 : 194240 - 194250
  • [3] Vehicle Speed Control by a Robotic Driver Considering Vehicle Dynamics for Continuously Variable Transmissions
    Mizutani, Naoto
    Matsui, Hirokazu
    Yano, Ken'ichi
    Takahashi, Toshimichi
    JOURNAL OF ROBOTICS AND MECHATRONICS, 2018, 30 (02) : 300 - 310
  • [4] An Improved Deep Deterministic Policy Gradient Pantograph Active Control Strategy for High-Speed Railways
    Wang, Ying
    Wang, Yuting
    Chen, Xiaoqiang
    Wang, Yixuan
    Chang, Zhanning
    ELECTRONICS, 2024, 13 (17)
  • [5] Deep Deterministic Policy Gradient With Compatible Critic Network
    Wang, Di
    Hu, Mengqi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4332 - 4344
  • [6] Robotic Visual-Inertial Calibration via Deep Deterministic Policy Gradient Learning
    Zhu, Wenxing
    Wang, Lihui
    Chen, Liangliang
    Xu, Ninghui
    Su, Yuzuwei
    IEEE SENSORS JOURNAL, 2022, 22 (14) : 14448 - 14457
  • [7] A Deep Deterministic Policy Gradient Learning Approach to Missile Autopilot Design
    Candeli, Angelo
    de Tommasi, Gianmaria
    Lui, Dario Giuseppe
    Mele, Adriano
    Santini, Stefania
    Tartaglione, Gaetano
    IEEE ACCESS, 2022, 10 : 19685 - 19696
  • [8] Ensemble Bootstrapped Deep Deterministic Policy Gradient for Vision-Based Robotic Grasping
    Liu, Weiwei
    Peng, Linpeng
    Cao, Junjie
    Fu, Xiaokuan
    Liu, Yong
    Pan, Zaisheng
    Yang, Jian
    IEEE ACCESS, 2021, 9 : 19916 - 19925
  • [9] Semicentralized Deep Deterministic Policy Gradient in Cooperative StarCraft Games
    Xie, Dong
    Zhong, Xiangnan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (04) : 1584 - 1593
  • [10] Grasping Control of a Vision Robot Based on a Deep Attentive Deterministic Policy Gradient
    Ji, Xiangxin
    Xiong, Feng
    Kong, Weichang
    Wei, Dongfei
    Shen, Zeyan
    IEEE ACCESS, 2022, 10 : 867 - 878