A Follow-the-Leader Strategy Using Hierarchical Deep Neural Networks with Grouped Convolutions

被引:0
作者
Solomon J.E. [1 ]
Charette F. [2 ]
机构
[1] Graphcore Inc., 167 Hamilton Ave, Palo Alto, 94301, CA
[2] Ford Motor Company, 3251 Hillview Ave, Palo Alto, 94304, CA
关键词
Autonomous systems; Computer vision; Deep learning; Grouped convolutions; Robotics;
D O I
10.1007/s42979-021-00572-1
中图分类号
学科分类号
摘要
The task of following-the-leader is implemented using a hierarchical deep neural network (DNN) end-to-end driving model to match the direction and speed of a target pedestrian. The model uses a classifier DNN to determine if the pedestrian is within the field of view of the camera sensor. If the pedestrian is present, the image stream from the camera is fed to a regression DNN which simultaneously adjusts the autonomous vehicle’s steering and throttle to keep cadence with the pedestrian. If the pedestrian is not visible, the vehicle uses a straightforward exploratory search strategy to reacquire the tracking objective. The classifier and regression DNNs incorporate grouped convolutions to boost model performance as well as to significantly reduce parameter count and compute latency. The models are trained on the intelligence processing unit (IPU) to leverage its fine-grain compute capabilities to minimize time-to-train. The results indicate very robust tracking behavior on the part of the autonomous vehicle in terms of its steering and throttle profiles, while requiring minimal data collection to produce. The throughput in terms of processing training samples has been boosted by the use of the IPU in conjunction with grouped convolutions by a factor ∼3.5 for training of the classifier and a factor of ∼7 for the regression network. A recording of the vehicle tracking a pedestrian has been produced and is available on the web. © 2021, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
引用
收藏
相关论文
共 16 条
  • [1] Baarath K., Zakaria M.A., Suparmaniam M., MYb A., Platooning strategy of mobile robot: simulation and experiment, MATEC Web Conf, 90, (2017)
  • [2] Zieba K. End to end learning for self-driving cars, 2016. Arxiv, (1604)
  • [3] Chowdhuri S., Pankaj T., Zipser K., Multi-Modal Multi-Task Deep Learning for Autonomous Driving. Corr., (2017)
  • [4] Goodchild A., Toy J., Delivery by drone: an evaluation of unmanned aerial vehicle technology in reducing CO <sub>2</sub> emissions in the delivery service industry, Transp Res Part D Transp Environ, (2017)
  • [5] Hierarchical Data Format Version 5., (2010)
  • [6] Iandola F.N., Moskewicz M.W., Ashraf K., Han S., Dally W.J., Keutzer K., Squeezenet: Alexnet-level accuracy with 50x fewer parameters and < 1 m b model size, Corr., (2016)
  • [7] Iandola F.N., Shaw A.E., Krishna R., Keutzer K.W., Squeezebert: What can computer vision teach nlp about efficient neural networks?, Corr., (2020)
  • [8] Ioffe S., Szegedy C., Batch normalization: Accelerating deep network training by reducing internal covariate shift, . Arxiv, (2015)
  • [9] Jia Z., Tillman B., Maggioni M., Scarpazza. D.P., Dissecting the graphcore ipu architecture via microbenchmarking, Arxiv, (2019)
  • [10] Klancar G., Matko D., Blazic S., Wheeled mobile robots control in a linear platoon, J Intell Robot Syst, 54, pp. 709-731, (2009)