Safe, human-like, decision-making for autonomous driving
Licentiate thesis, 2022
We formulate the tactical decision-making task in driving as a sequential decision-making problem and describe it with Markov decision processes (MDP). Reinforcement learning (RL) techniques are adopted as the backbone in solving the MDP. To ensure safety, we propose a system architecture to combine shielding with RL. Shielding is a formal method to prevent learning methods from taking dangerous actions. Furthermore, the human driving experience is used to improve the data efficiency for RL methods and make the driving policy more human-like.
Although RL methods can solve decision-making problems, the performance heavily depends on reward functions. Since the true reward function for driving is unknown, we address this problem using imitation learning. Adversarial inverse reinforcement learning (AIRL) extracts both the reward function and the driving policy from expert driving demonstrations. To improve and stabilize the performance, we propose reward augmentations for AIRL and demonstrate better results.
human-like behaviors.
imitation learning
Autonomous driving
reinforcement learning
decision making
safety
Author
Dapeng Liu
Chalmers, Electrical Engineering, Communication, Antennas and Optical Networks
Decision Making for Autonomous Driving via Augmented Adversarial Inverse Reinforcement Learning
Proceedings - IEEE International Conference on Robotics and Automation,;Vol. 2021-May(2021)p. 1036-1042
Paper in proceeding
Learning faster to perform autonomous lane changes by constructing maneuvers from shielded semantic actions
2019 IEEE Intelligent Transportation Systems Conference, ITSC 2019,;(2019)p. 1838-1844
Paper in proceeding
Areas of Advance
Information and Communication Technology
Transport
Subject Categories
Information Science
Computer Science
Computer Systems
Publisher
Chalmers
Online participation via Zoom and room EA, floor 4, Hörsalsvägen 11
Opponent: Dr. Wei Zhan, University of California, Berkeley