This content is not included in your SAE MOBILUS subscription, or you are not logged in.
Heterogeneous Machine Learning on High Performance Computing for End to End Driving of Autonomous Vehicles
ISSN: 0148-7191, e-ISSN: 2688-3627
Published April 14, 2020 by SAE International in United States
Annotation ability available
Current artificial intelligence techniques for end to end driving of autonomous vehicles typically rely on a single form of learning or training processes along with a corresponding dataset or simulation environment. Relatively speaking, success has been shown for a variety of learning modalities in which it can be shown that the machine can successfully “drive” a vehicle. However, the realm of real-world driving extends significantly beyond the realm of limited test environments for machine training. This creates an enormous gap in capability between these two realms. With their superior neural network structures and learning capabilities, humans can be easily trained within a short period of time to proceed from limited test environments to real world driving. For machines though, this gap is guarded by at least two challenges: 1) machine learning techniques remain brittle and unable to generalize to a wide range of scenarios, and 2) effective training data that enhances generalization and generates the desired driving behavior. Further, each challenge can be computationally intensive on its own thereby exasperating the gap. Moreover, is has not yet been shown that a single form of learning or training is capable of addressing a large range of scenarios. As a result, solving the first challenge does not inherently solve the second and vice versa. The work described here discusses an approach to address the first challenge that would also provide a foundation for solving the second. Our approach utilizes a combination of conditional imitation learning with a static dataset, reinforcement learning with a simulation environment, and high-performance computing to train a neural network. As a result, this reduces the “time to solution” from to the existing techniques for autonomous driving and provides an extensible framework to address the second key challenge.
- Robert Patton - Oak Ridge National Laboratory
- Shang Gao - Oak Ridge National Laboratory
- Spencer Paulissen - Oak Ridge National Laboratory
- Nicholas Haas - Oak Ridge National Laboratory
- Brian Jewell - Oak Ridge National Laboratory
- Xiangyu Zhang - National Renewable Energy Laboratory
- Peter Graf - National Renewable Energy Laboratory
CitationPatton, R., Gao, S., Paulissen, S., Haas, N. et al., "Heterogeneous Machine Learning on High Performance Computing for End to End Driving of Autonomous Vehicles," SAE Technical Paper 2020-01-0739, 2020, https://doi.org/10.4271/2020-01-0739.
- Pomerleau, D.A. , “ALVINN: An Autonomous Land Vehicle in a Neural Network,” Advances in Neural Information Processing Systems (NIPS), 1988.
- Canziani, A., Paszke, A., and Culurciello, E. , “An Analysis of Deep Neural Network Models for Practical Applications,” arXiv preprint arXiv:1605.07678, 2016.
- Turing, A. , “Computing Machinery and Intelligence,” Mind 59:433-460, 1950.
- Rosenblatt, F. , “The Perceptron: A Probabilistic Model for Information Storage and Organization in the Brain,” Psychological Review 65(6):386-408, 1958, doi:10.1037/h0042519.
- Samuel, A.L. , “Some Studies in Machine Learning Using the Game of Checkers,” IBM Journal of Research and Development 44(1.2):206-226, Jan. 2000, doi:10.1147/rd.441.0206.
- LeCun, Y., Bengio, Y., and Hinton, G. , “Deep Learning,” Nature 521:436, 2015.
- Osa, T. et al. , “An Algorithmic Perspective on Imitation Learning”, arXiv:1811.06711, Nov 2018.
- Kaelbling, L.P., Littman, M.L., and Moore, A.W. , “Reinforcement Learning: A Survey,” Journal of Artificial Intelligence Research 4:237-285, 1996, doi:10.1613/jair.301.
- François-Lavet, V., Henderson, P., Islam, R., Bellemare, M.G., and Pineau, J. , “An Introduction to Deep Reinforcement Learnig,” Foundations and Trends in Machine Learning 11(3-4):219-354, 2018, http://dx.doi.org/10.1561/2200000071.
- Dosovitskiy, A., Ros, G., Codevilla, F., Lopez, A., and Koltun, V. , “CARLA: An Open Urban Driving Simulator,” in Conference on Robot Learning, 2017, 1-16.
- Bojarski et al. , “End to End Learning for Self-Driving Cars,” arXiv:1604.07316, April 2016.
- Codevilla et al. , “End-to-End Driving via Conditional Imitation Learning,” arXiv:1710.02410, Oct 2017.
- Rhinehart et al. , “Deep Imitative Models for Flexible Inference, Planning, and Control,” arXiv:1810.06544, Oct 2018.
- Bansal, M., Krizhevsky, A., and Ogale, A. , “ChauffeurNet: Learning to Drive by Imitating the Best and Synthesizing the Worst,” arXiv:1812.03079, Dec 2018.
- Okuyama, T., Gonsalves, T., and Upadhay, J. “Autonomous Driving System Based on Deep Q Learnig,” in 2018 International Conference on Intelligent Autonomous Systems (ICoIAS), Singapore, 2018, 201-205, doi:10.1109/ICoIAS.2018.8494053.
- Kendall, A. et al. , “Learning to Drive in a Day,” arXiv:1807.00412, Jul 2018.
- Cruz, G., Du, Y., and Taylor, M.E. , “Pre-Training with Non-Expert Human Demonstration for Deep Reinforcement Learning,” arXiv:1812.08904, Dec 2018.
- Liang, X., Wang, T., Yang, L., and Xing, E. , “CIRL: Controllable Imitative Reinforcement Learning for Vision-based Self-Driving”, arXiv:1807.03776, July 2018.
- Sutton, R.S., and Barto, A.G. (1998). Introduction to Reinforcement Learning, Vol. 2, No. 4 (Cambridge: MIT Press).
- Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and Klimov, O. , “Proximal Policy Optimization Algorithms,” arXiv:1707.06347, July 2017.
- Schulman, J., Levine, S., Abbeel, P., Jordan, M., and Moritz, P. , “Trust Region Policy Optimization,” in Proceedings of the 32nd International Conference on Machine Learning, PMLR 37, June 2015, 1889-1897.
- Brockman, G. et al. , “OpenAI Gym,” arXiv:1606.01540, June 2016.
- “CARLA Cityscapes Color Palette,” https://github.com/carla-simulator/carla/blob/46d36435d4c0b769a5de20077739e9490e50e5ad/LibCarla/source/carla/image/CityScapesPalette.h#L19-L52, accessed November 2019.
- “CARLA Gym Environment Based on Multiple CARLA Servers,” https://github.com/ray-project/ray/blob/releases/0.7.0/python/ray/rllib/examples/carla/env.py#L186, accessed November 2019.