Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles

Automated driving is essential for developing and deploying intelligent transportation systems. However, unavoidable sensor noises or perception errors may cause an automated vehicle to adopt suboptimal driving policies or even lead to catastrophic failures. Additionally, the automated driving longi...

Full description

Saved in:
Bibliographic Details
Main Authors: He, Xiangkun, Chen, Hao, Lv, Chen
Other Authors: School of Mechanical and Aerospace Engineering
Format: Article
Language:English
Published: 2024
Subjects:
Online Access:https://hdl.handle.net/10356/173497
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-173497
record_format dspace
spelling sg-ntu-dr.10356-1734972024-02-10T16:47:58Z Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles He, Xiangkun Chen, Hao Lv, Chen School of Mechanical and Aerospace Engineering Engineering Automated Vehicle Vehicle Dynamics Automated driving is essential for developing and deploying intelligent transportation systems. However, unavoidable sensor noises or perception errors may cause an automated vehicle to adopt suboptimal driving policies or even lead to catastrophic failures. Additionally, the automated driving longitudinal and lateral decision-making behaviors (e.g., driving speed and lane changing decisions) are coupled, that is, when one of them is perturbed by unknown external disturbances, it causes changes or even performance degradation in the other. The presence of both challenges significantly curtails the potential of automated driving. Here, to coordinate the longitudinal and lateral driving decisions of an automated vehicle while ensuring policy robustness against observational uncertainties, we propose a novel robust coordinated decision-making technique via robust multiagent reinforcement learning. Specifically, the automated driving longitudinal and lateral decisions under observational perturbations are modeled as a constrained robust multiagent Markov decision process. Meanwhile, a nonlinear constraint setting with Kullback-Leibler divergence is developed to keep the variation of the driving policy perturbed by stochastic perturbations within bounds. Additionally, a robust multiagent policy optimization approach is proposed to approximate the optimal robust coordinated driving policy. Finally, we evaluate the proposed robust coordinated decision-making method in three highway scenarios with different traffic densities. Quantitatively, in the absence of noises, the proposed method achieves an approximate average enhancement of 25.58% in traffic efficiency and 91.31% in safety compared to all baselines across the three scenarios. In the presence of noises, our technique improves traffic efficiency and safety by an approximate average of 30.81% and 81.02% compared to all baselines in the three scenarios, respectively. The results demonstrate that the proposed approach is capable of improving automated driving performance and ensuring policy robustness against observational uncertainties. Nanyang Technological University Published version This work was supported by Strat-Up Grant of Nanyang Technological University. 2024-02-07T05:08:18Z 2024-02-07T05:08:18Z 2023 Journal Article He, X., Chen, H. & Lv, C. (2023). Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles. SAE International Journal of Vehicle Dynamics, Stability, and NVH, 7(4), 475-488. https://dx.doi.org/10.4271/10-07-04-0031 2380-2162 https://hdl.handle.net/10356/173497 10.4271/10-07-04-0031 2-s2.0-85173050021 4 7 475 488 en SAE International Journal of Vehicle Dynamics, Stability, and NVH © 2023 Nanyang Technological University; Published by SAE International. This Open Access article is published under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/), which permits distribution, and reproduction in any medium, provided that the original author(s) and the source are credited. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Automated Vehicle
Vehicle Dynamics
spellingShingle Engineering
Automated Vehicle
Vehicle Dynamics
He, Xiangkun
Chen, Hao
Lv, Chen
Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
description Automated driving is essential for developing and deploying intelligent transportation systems. However, unavoidable sensor noises or perception errors may cause an automated vehicle to adopt suboptimal driving policies or even lead to catastrophic failures. Additionally, the automated driving longitudinal and lateral decision-making behaviors (e.g., driving speed and lane changing decisions) are coupled, that is, when one of them is perturbed by unknown external disturbances, it causes changes or even performance degradation in the other. The presence of both challenges significantly curtails the potential of automated driving. Here, to coordinate the longitudinal and lateral driving decisions of an automated vehicle while ensuring policy robustness against observational uncertainties, we propose a novel robust coordinated decision-making technique via robust multiagent reinforcement learning. Specifically, the automated driving longitudinal and lateral decisions under observational perturbations are modeled as a constrained robust multiagent Markov decision process. Meanwhile, a nonlinear constraint setting with Kullback-Leibler divergence is developed to keep the variation of the driving policy perturbed by stochastic perturbations within bounds. Additionally, a robust multiagent policy optimization approach is proposed to approximate the optimal robust coordinated driving policy. Finally, we evaluate the proposed robust coordinated decision-making method in three highway scenarios with different traffic densities. Quantitatively, in the absence of noises, the proposed method achieves an approximate average enhancement of 25.58% in traffic efficiency and 91.31% in safety compared to all baselines across the three scenarios. In the presence of noises, our technique improves traffic efficiency and safety by an approximate average of 30.81% and 81.02% compared to all baselines in the three scenarios, respectively. The results demonstrate that the proposed approach is capable of improving automated driving performance and ensuring policy robustness against observational uncertainties.
author2 School of Mechanical and Aerospace Engineering
author_facet School of Mechanical and Aerospace Engineering
He, Xiangkun
Chen, Hao
Lv, Chen
format Article
author He, Xiangkun
Chen, Hao
Lv, Chen
author_sort He, Xiangkun
title Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
title_short Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
title_full Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
title_fullStr Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
title_full_unstemmed Robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
title_sort robust multiagent reinforcement learning toward coordinated decision-making of automated vehicles
publishDate 2024
url https://hdl.handle.net/10356/173497
_version_ 1794549285126144000