A new reinforcement learning vehicle control architecture for vision-based road following

被引:47
作者
Oh, SY [1 ]
Lee, JH
Choi, DH
机构
[1] Pohang Univ Sci & Technol, Dept Elect Engn, Pohang 790784, South Korea
[2] Penta Secur Syst Inc, Penta Secur Technol Lab, Seoul, South Korea
[3] Seoul Natl Univ, Coll Engn, Sch Elect Engn & Comp Sci, Seoul, South Korea
关键词
lateral control; neural networks; reinforcement learning; road following; vehicle dynamics;
D O I
10.1109/25.845116
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
A new dynamic control architecture based on reinforcement learning (RL) has been developed and applied to the problem of high-speed road following of high-curvature roads. Through RL, the control system indirectly learns the vehicle-road interaction dynamics, knowledge which is essential to stay on the road in high speed road tracking. First, computer simulation has been carried out in order to test stability and performance of the proposed RL controller before actual use. The proposed controller exhibited a good road tracking performance, especially on high-curvature roads, Then, the actual autonomous driving experiments successfully verified the control performance on campus roads in which there were shadows from the trees, noisy and/or broken lane markings, different road curvatures, and also different times of the day reflecting a range of lighting conditions. The proposed three-stage image processing algorithm and the use of all six strips of edges have been capable of handling most of the uncertainties arising from the nonideal road conditions.
引用
收藏
页码:997 / 1005
页数:9
相关论文
共 19 条
[1]  
BARTO A, 1983, IEEE T SYST MAN CYBE, V13
[2]   PARALLEL AND LOCAL FEATURE-EXTRACTION - A REAL-TIME APPROACH TO ROAD BOUNDARY DETECTION [J].
BROGGI, A .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 1995, 4 (02) :217-223
[3]  
CHOI DH, 1996, NEURAL PARALLEL SCI, V4, P367
[4]   RECURSIVE 3-D ROAD AND RELATIVE EGO-STATE RECOGNITION [J].
DICKMANNS, ED ;
MYSLIWETZ, BD .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1992, 14 (02) :199-213
[5]   A STOCHASTIC REINFORCEMENT LEARNING ALGORITHM FOR LEARNING REAL-VALUED FUNCTIONS [J].
GULLAPALLI, V .
NEURAL NETWORKS, 1990, 3 (06) :671-692
[6]  
Haralick R. M., 1992, COMPUTER ROBOT VISIO, V1
[7]  
HEDRICK JK, 1994, IEEE CONTROL SYS DEC, P21
[8]  
KAELBLING LP, 1986, J ARTIFICIAL INTELL, P237
[9]  
Lee JH, 1998, IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, P2028, DOI 10.1109/IJCNN.1998.687171
[10]  
Lin CF, 1998, J ROBOTIC SYST, V15, P537, DOI 10.1002/(SICI)1097-4563(199810)15:10<537::AID-ROB1>3.0.CO