Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach

As a strategy to reduce travel delay and enhance energy efficiency, platooning of connected and autonomous vehicles (CAVs) at non-signalized intersections has become increasingly popular in academia. However, few studies have attempted to model the relation between the optimal platoon size and the t...

وصف كامل

محفوظ في:
التفاصيل البيبلوغرافية
المؤلفون الرئيسيون: Li, Duowei, Wu, Jianping, Zhu, Feng, Chen, Tianyi, Wong, Yiik Diew
مؤلفون آخرون: School of Civil and Environmental Engineering
التنسيق: Conference or Workshop Item
اللغة:English
منشور في: 2022
الموضوعات:
الوصول للمادة أونلاين:https://hdl.handle.net/10356/158101
https://heart2022.com/
الوسوم: إضافة وسم
لا توجد وسوم, كن أول من يضع وسما على هذه التسجيلة!
id sg-ntu-dr.10356-158101
record_format dspace
spelling sg-ntu-dr.10356-1581012022-07-05T07:22:22Z Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach Li, Duowei Wu, Jianping Zhu, Feng Chen, Tianyi Wong, Yiik Diew School of Civil and Environmental Engineering hEART 2022: 10th Symposium of the European Association for Research in Transportation Engineering::Civil engineering::Transportation Engineering::Computer science and engineering::Computer applications Connected and Autonomous Vehicle Adaptive Platoon Intersection Control Deep Reinforcement Learning As a strategy to reduce travel delay and enhance energy efficiency, platooning of connected and autonomous vehicles (CAVs) at non-signalized intersections has become increasingly popular in academia. However, few studies have attempted to model the relation between the optimal platoon size and the traffic conditions around the intersection. To this end, this study proposes an adaptive platoon based autonomous intersection control model powered by deep reinforcement learning (DRL) technique. The model framework has following two levels: the first level adopts a First Come First Serve (FCFS) reservation based policy integrated with a nonconflicting lane selection mechanism to determine vehicles’ passing priority; and the second level applies a deep Q-network algorithm to identify the optimal platoon size based on the real-time traffic condition of an intersection. When being tested on a traffic micro-simulator, our proposed model exhibits superior performances on travel efficiency and fuel conservation as compared to the state-of-the-art methods. Submitted/Accepted version 2022-07-05T05:30:59Z 2022-07-05T05:30:59Z 2022 Conference Paper Li, D., Wu, J., Zhu, F., Chen, T. & Wong, Y. D. (2022). Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach. hEART 2022: 10th Symposium of the European Association for Research in Transportation, 1-11. https://hdl.handle.net/10356/158101 https://heart2022.com/ 1 11 en © 2022 The Author(s). All rights reserved. This paper was published in Proceedings of hEART 2022: 10th Symposium of the European Association for Research in Transportation and is made available with permission of The Author(s). application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Civil engineering::Transportation
Engineering::Computer science and engineering::Computer applications
Connected and Autonomous Vehicle
Adaptive Platoon
Intersection Control
Deep Reinforcement Learning
spellingShingle Engineering::Civil engineering::Transportation
Engineering::Computer science and engineering::Computer applications
Connected and Autonomous Vehicle
Adaptive Platoon
Intersection Control
Deep Reinforcement Learning
Li, Duowei
Wu, Jianping
Zhu, Feng
Chen, Tianyi
Wong, Yiik Diew
Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
description As a strategy to reduce travel delay and enhance energy efficiency, platooning of connected and autonomous vehicles (CAVs) at non-signalized intersections has become increasingly popular in academia. However, few studies have attempted to model the relation between the optimal platoon size and the traffic conditions around the intersection. To this end, this study proposes an adaptive platoon based autonomous intersection control model powered by deep reinforcement learning (DRL) technique. The model framework has following two levels: the first level adopts a First Come First Serve (FCFS) reservation based policy integrated with a nonconflicting lane selection mechanism to determine vehicles’ passing priority; and the second level applies a deep Q-network algorithm to identify the optimal platoon size based on the real-time traffic condition of an intersection. When being tested on a traffic micro-simulator, our proposed model exhibits superior performances on travel efficiency and fuel conservation as compared to the state-of-the-art methods.
author2 School of Civil and Environmental Engineering
author_facet School of Civil and Environmental Engineering
Li, Duowei
Wu, Jianping
Zhu, Feng
Chen, Tianyi
Wong, Yiik Diew
format Conference or Workshop Item
author Li, Duowei
Wu, Jianping
Zhu, Feng
Chen, Tianyi
Wong, Yiik Diew
author_sort Li, Duowei
title Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
title_short Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
title_full Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
title_fullStr Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
title_full_unstemmed Modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
title_sort modeling adaptive platoon and reservation based autonomous intersection control: a deep reinforcement learning approach
publishDate 2022
url https://hdl.handle.net/10356/158101
https://heart2022.com/
_version_ 1738844960706265088