Towards robust car-following based on deep reinforcement learning
Research output: Contribution to journal › Research article › Contributed › peer-review
Contributors
Abstract
One of the biggest challenges in the development of learning-driven automated driving technologies remains the handling of uncommon, rare events that may have not been encountered in training. Especially when training a model with real driving data, unusual situations, such as emergency brakings, may be underrepresented, resulting in a model that lacks robustness in rare events. This study focuses on car-following based on reinforcement learning and demonstrates that existing approaches, trained with real driving data, fail to handle safety–critical situations. Since collecting data representing all kinds of possible car-following events, including safety–critical situations, is challenging, we propose a training environment that harnesses stochastic processes to generate diverse and challenging scenarios. Our experiments show that training with real data can lead to models that collide in safety–critical situations, whereas the proposed model exhibits excellent performance and remains accident-free, comfortable, and string-stable even in extreme scenarios, such as full-braking by the leading vehicle. Its robustness is demonstrated by simulating car-following scenarios for various reward function parametrizations and a diverse range of artificial and real leader data that were not included in training and were qualitatively different from the learning data. We further show that conventional reward designs can encourage aggressive behavior when approaching other vehicles. Additionally, we compared the proposed model with classical car-following models and found it to achieve equal or superior results.
Details
Original language | English |
---|---|
Article number | 104486 |
Journal | Transportation Research Part C: Emerging Technologies |
Volume | 159 |
Publication status | Published - Feb 2024 |
Peer-reviewed | Yes |
External IDs
ORCID | /0000-0002-8909-4861/work/171064872 |
---|---|
ORCID | /0000-0002-1730-0750/work/171065331 |
Keywords
ASJC Scopus subject areas
Keywords
- Car-following model, Generalization capabilities, Reinforcement learning, String stability, Validation