Safe, human-like, decision-making for autonomous driving
Licentiatavhandling, 2022
We formulate the tactical decision-making task in driving as a sequential decision-making problem and describe it with Markov decision processes (MDP). Reinforcement learning (RL) techniques are adopted as the backbone in solving the MDP. To ensure safety, we propose a system architecture to combine shielding with RL. Shielding is a formal method to prevent learning methods from taking dangerous actions. Furthermore, the human driving experience is used to improve the data efficiency for RL methods and make the driving policy more human-like.
Although RL methods can solve decision-making problems, the performance heavily depends on reward functions. Since the true reward function for driving is unknown, we address this problem using imitation learning. Adversarial inverse reinforcement learning (AIRL) extracts both the reward function and the driving policy from expert driving demonstrations. To improve and stabilize the performance, we propose reward augmentations for AIRL and demonstrate better results.
human-like behaviors.
imitation learning
Autonomous driving
reinforcement learning
decision making
safety
Författare
Dapeng Liu
Chalmers, Elektroteknik, Kommunikation, Antenner och Optiska Nätverk
Decision Making for Autonomous Driving via Augmented Adversarial Inverse Reinforcement Learning
Proceedings - IEEE International Conference on Robotics and Automation,;Vol. 2021-May(2021)p. 1036-1042
Paper i proceeding
Learning faster to perform autonomous lane changes by constructing maneuvers from shielded semantic actions
2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019,;(2019)p. 1838-1844
Paper i proceeding
Styrkeområden
Informations- och kommunikationsteknik
Transport
Ämneskategorier
Systemvetenskap
Datavetenskap (datalogi)
Datorsystem
Utgivare
Chalmers
Online participation via Zoom and room EA, floor 4, Hörsalsvägen 11
Opponent: Dr. Wei Zhan, University of California, Berkeley