Learning from Simulation, Racing in Reality: Sim2Real Methods for Autonomous Racing

Reinforcement Learning (RL) methods have been successfully demonstrated in robotic tasks, however, their application to continuous state-action systems with fast dynamics is challenging. In this work, we investigate RL solutions for the autonomous racing problem on the ORCA miniature race car platfo...

Full description

Bibliographic Details
Main Author: Chisari, Eugenio
Other Authors: Rupenyan, Alisa, Liniger, Alexander, Balula, Samuel, Lygeros, John
Language:English
Published: ETH Zurich 2020
Subjects:
Online Access:https://hdl.handle.net/20.500.11850/486540
https://doi.org/10.3929/ethz-b-000486540
Description
Summary:Reinforcement Learning (RL) methods have been successfully demonstrated in robotic tasks, however, their application to continuous state-action systems with fast dynamics is challenging. In this work, we investigate RL solutions for the autonomous racing problem on the ORCA miniature race car platform. When training a deep neural network policy using RL methods only using simulations, we observe poor performance, due to model mismatch also known as reality gap. We propose three different methods to reduce this gap, first we propose a policy regularization in the policy optimization step, second, we use model randomization. These two methods allow learning a policy that can race the car without any real environment interactions. Our third method improves this policy, by running the RL algorithm online while driving the car. The achieved performance on the ORCA platform is comparable to that achieved previously by a model-based controller, in terms of lap time, and improved with respect to track constraint violations.