TY - GEN
T1 - IVDR
T2 - 20th IEEE International Conference on Machine Learning and Applications, ICMLA 2021
AU - Joo, Kihyung
AU - Woo, Simon S.
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Current state-of-the-art autonomous driving technology significantly advanced, leveraging reinforcement learning (RL) algorithms, because it is not easy to apply a rule-based driving method that reflects all the various traffic conditions. Indeed, reinforcement learning can produce the possible optimal driving strategy of urban, rural, and motorway roads in various environmental conditions such as speed limits and school zones. However, it is challenging to adjust the parameters of the reward mechanism in RL, because the driving style of each user is very different. And it takes a massive amount of time and resources to conduct RL by reflecting all complex traffic conditions. However, if RL imitates the driving behavior of an expert, RL algorithm can proceed more quickly. Therefore, we propose a novel imitation learning framework, which combines an expert's driving behavior with a continuous behavior of an agent. Further, a deep reinforcement learning approach is used to mimic the expert's driving behavior. Therefore, we propose imitation learning with variational inference and distributional reinforcement learning (IVDR) algorithm. Our results show that IVDR achieves 80% better learning speed than the learning speed of other approaches and outperforms 12% higher in average reward. Our work shows great promise of using RL for autonomous driving and real vehicle driving simulation.
AB - Current state-of-the-art autonomous driving technology significantly advanced, leveraging reinforcement learning (RL) algorithms, because it is not easy to apply a rule-based driving method that reflects all the various traffic conditions. Indeed, reinforcement learning can produce the possible optimal driving strategy of urban, rural, and motorway roads in various environmental conditions such as speed limits and school zones. However, it is challenging to adjust the parameters of the reward mechanism in RL, because the driving style of each user is very different. And it takes a massive amount of time and resources to conduct RL by reflecting all complex traffic conditions. However, if RL imitates the driving behavior of an expert, RL algorithm can proceed more quickly. Therefore, we propose a novel imitation learning framework, which combines an expert's driving behavior with a continuous behavior of an agent. Further, a deep reinforcement learning approach is used to mimic the expert's driving behavior. Therefore, we propose imitation learning with variational inference and distributional reinforcement learning (IVDR) algorithm. Our results show that IVDR achieves 80% better learning speed than the learning speed of other approaches and outperforms 12% higher in average reward. Our work shows great promise of using RL for autonomous driving and real vehicle driving simulation.
KW - Distributional RL (DR)
KW - Imitation learning
KW - Reinforcement Learning (RL)
KW - Soft Actor-Critic (SAC)
KW - Variational Inference (VI)
UR - https://www.scopus.com/pages/publications/85125856027
U2 - 10.1109/ICMLA52953.2021.00047
DO - 10.1109/ICMLA52953.2021.00047
M3 - Conference contribution
AN - SCOPUS:85125856027
T3 - Proceedings - 20th IEEE International Conference on Machine Learning and Applications, ICMLA 2021
SP - 256
EP - 262
BT - Proceedings - 20th IEEE International Conference on Machine Learning and Applications, ICMLA 2021
A2 - Wani, M. Arif
A2 - Sethi, Ishwar K.
A2 - Shi, Weisong
A2 - Qu, Guangzhi
A2 - Raicu, Daniela Stan
A2 - Jin, Ruoming
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 13 December 2021 through 16 December 2021
ER -