Not logged in.
Quick Search - Contribution
Contribution Details
Type | Journal Article |
Scope | Discipline-based scholarship |
Title | Super-Human Performance in Gran Turismo Sport Using Deep Reinforcement Learning |
Organization Unit | |
Authors |
|
Item Subtype | Original Work |
Refereed | Yes |
Status | Published in final form |
Language |
|
Journal Title | IEEE Robotics and Automation Letters |
Publisher | Institute of Electrical and Electronics Engineers |
Geographical Reach | international |
ISSN | 2377-3766 |
Volume | 6 |
Number | 3 |
Page Range | 4257 - 4264 |
Date | 2022 |
Abstract Text | Autonomous car racing is a major challenge in robotics. It raises fundamental problems for classical approaches such as planning minimum-time trajectories under uncertain dynamics and controlling the car at the limits of its handling. Besides, the requirement of minimizing the lap time, which is a sparse objective, and the difficulty of collecting training data from human experts have also hindered researchers from directly applying learning-based approaches to solve the problem. In the present work, we propose a learning-based system for autonomous car racing by leveraging a high-fidelity physical car simulation, a course-progress proxy reward, and deep reinforcement learning. We deploy our system in Gran Turismo Sport, a world-leading car simulator known for its realistic physics simulation of different race cars and tracks, which is even used to recruit human race car drivers. Our trained policy achieves autonomous racing performance that goes beyond what had been achieved so far by the built-in AI, and at the same time, outperforms the fastest driver in a dataset of over 50,000 human players. |
Digital Object Identifier | 10.1109/LRA.2021.3064284 |
Other Identification Number | merlin-id:22159 |
PDF File | Download from ZORA |
Export |
BibTeX
EP3 XML (ZORA) |