oai:arXiv.org:2406.08878
Computer Science
2024
11/13/2024
Modern approaches to autonomous driving rely heavily on learned components trained with large amounts of human driving data via imitation learning.
However, these methods require large amounts of expensive data collection and even then face challenges with safely handling long-tail scenarios and compounding errors over time.
At the same time, pure Reinforcement Learning (RL) methods can fail to learn performant policies in sparse, constrained, and challenging-to-define reward settings such as autonomous driving.
Both of these challenges make deploying purely cloned or pure RL policies in safety critical applications such as autonomous vehicles challenging.
In this paper we propose Combining IMitation and Reinforcement Learning (CIMRL) approach - a safe reinforcement learning framework that enables training driving policies in simulation through leveraging imitative motion priors and safety constraints.
CIMRL does not require extensive reward specification and improves on the closed loop behavior of pure cloning methods.
By combining RL and imitation, we demonstrate that our method achieves state-of-the-art results in closed loop simulation and real world driving benchmarks.
Booher, Jonathan,Rohanimanesh, Khashayar,Xu, Junhong,Isenbaev, Vladislav,Balakrishna, Ashwin,Gupta, Ishan,Liu, Wei,Petiushko, Aleksandr, 2024, CIMRL: Combining IMitation and Reinforcement Learning for Safe Autonomous Driving