Abstract
Reinforcement learning (RL) is an effective approach to motion planning in autonomous driving, where an optimal driving policy can be automatically learned using the interaction data with the environment. Nevertheless, the reward function for an RL agent, which is significant to its performance, is challenging to determine. The conventional work mainly focuses on rewarding safe driving states but does not incorporate the awareness of risky driving behaviors of the vehicles. In this paper, we investigate how to use risk-aware reward shaping to leverage the training and test performance of RL agents in autonomous driving. Based on the essential requirements that prescribe the safety specifications for general autonomous driving in practice, we propose additional reshaped reward terms that encourage exploration and penalize risky driving behaviors. A simulation study in OpenAI Gym indicates the advantage of risk-aware reward shaping for various RL agents. Also, we point out that proximal policy optimization (PPO) is likely to be the best RL method that works with risk-aware reward shaping.
Original language | English |
---|---|
Title of host publication | IECON 2023 - 49th Annual Conference of the IEEE Industrial Electronics Society |
Publisher | Institute of Electrical and Electronics Engineers |
Number of pages | 6 |
ISBN (Electronic) | 979-8-3503-3182-0 |
DOIs | |
Publication status | Published - 16 Nov 2023 |
Event | 49th Annual Conference of the IEEE Industrial Electronics Society, IECON 2023 - Singapore, Singapore Duration: 16 Oct 2023 → 19 Oct 2023 |
Conference
Conference | 49th Annual Conference of the IEEE Industrial Electronics Society, IECON 2023 |
---|---|
Country/Territory | Singapore |
City | Singapore |
Period | 16/10/23 → 19/10/23 |
Funding
This work was supported by the European project SymAware under grant No. 101070802.
Keywords
- autonomous driving
- motion planning
- reinforcement learning
- reward shaping
- risk awareness