Using reinforcement learning to minimize the probability of delay occurrence in transportation
Reducing traffic delay is of crucial importance for the development of sustainable transportation systems, which is a challenging task in the studies of stochastic shortest path (SSP) problem. Existing methods based on the probability tail model to solve the SSP problem, seek for the path that minim...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2020
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8158 https://ink.library.smu.edu.sg/context/sis_research/article/9161/viewcontent/Using_Reinforcement_Learning_to_Minimize_the_Probability_of_Delay_Occurrence_in_Transportation__1_.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9161 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-91612023-09-26T10:38:56Z Using reinforcement learning to minimize the probability of delay occurrence in transportation CAO, Zhiguang Guo, Hongliang Song, Wen Gao, Kaizhou Chen, Zhengghua Zhang, Le Zhang, Xuexi Reducing traffic delay is of crucial importance for the development of sustainable transportation systems, which is a challenging task in the studies of stochastic shortest path (SSP) problem. Existing methods based on the probability tail model to solve the SSP problem, seek for the path that minimizes the probability of delay occurrence, which is equal to maximizing the probability of reaching the destination before a deadline (i.e., arriving on time). However, they suffer from low accuracy or high computational cost. Therefore, we design a novel and practical Q-learning approach where the converged Q-values have the practical meaning as the actual probabilities of arriving on time so as to improve the accuracy of finding the real optimal path. By further adopting dynamic neural networks to learn the value function, our approach can scale well to large road networks with arbitrary deadlines. Moreover, our approach is flexible to implement in a time dependent manner, which further improves the performance of returned path. Experimental results on some road networks with real mobility data, such as Beijing, Munich and Singapore, demonstrate the significant advantages of the proposed approach over other methods. 2020-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8158 info:doi/10.1109/TVT.2020.2964784 https://ink.library.smu.edu.sg/context/sis_research/article/9161/viewcontent/Using_Reinforcement_Learning_to_Minimize_the_Probability_of_Delay_Occurrence_in_Transportation__1_.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Reinforcement learning Transportation Arriving on time Vehicle routing Q-learning Databases and Information Systems Transportation |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Reinforcement learning Transportation Arriving on time Vehicle routing Q-learning Databases and Information Systems Transportation |
spellingShingle |
Reinforcement learning Transportation Arriving on time Vehicle routing Q-learning Databases and Information Systems Transportation CAO, Zhiguang Guo, Hongliang Song, Wen Gao, Kaizhou Chen, Zhengghua Zhang, Le Zhang, Xuexi Using reinforcement learning to minimize the probability of delay occurrence in transportation |
description |
Reducing traffic delay is of crucial importance for the development of sustainable transportation systems, which is a challenging task in the studies of stochastic shortest path (SSP) problem. Existing methods based on the probability tail model to solve the SSP problem, seek for the path that minimizes the probability of delay occurrence, which is equal to maximizing the probability of reaching the destination before a deadline (i.e., arriving on time). However, they suffer from low accuracy or high computational cost. Therefore, we design a novel and practical Q-learning approach where the converged Q-values have the practical meaning as the actual probabilities of arriving on time so as to improve the accuracy of finding the real optimal path. By further adopting dynamic neural networks to learn the value function, our approach can scale well to large road networks with arbitrary deadlines. Moreover, our approach is flexible to implement in a time dependent manner, which further improves the performance of returned path. Experimental results on some road networks with real mobility data, such as Beijing, Munich and Singapore, demonstrate the significant advantages of the proposed approach over other methods. |
format |
text |
author |
CAO, Zhiguang Guo, Hongliang Song, Wen Gao, Kaizhou Chen, Zhengghua Zhang, Le Zhang, Xuexi |
author_facet |
CAO, Zhiguang Guo, Hongliang Song, Wen Gao, Kaizhou Chen, Zhengghua Zhang, Le Zhang, Xuexi |
author_sort |
CAO, Zhiguang |
title |
Using reinforcement learning to minimize the probability of delay occurrence in transportation |
title_short |
Using reinforcement learning to minimize the probability of delay occurrence in transportation |
title_full |
Using reinforcement learning to minimize the probability of delay occurrence in transportation |
title_fullStr |
Using reinforcement learning to minimize the probability of delay occurrence in transportation |
title_full_unstemmed |
Using reinforcement learning to minimize the probability of delay occurrence in transportation |
title_sort |
using reinforcement learning to minimize the probability of delay occurrence in transportation |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2020 |
url |
https://ink.library.smu.edu.sg/sis_research/8158 https://ink.library.smu.edu.sg/context/sis_research/article/9161/viewcontent/Using_Reinforcement_Learning_to_Minimize_the_Probability_of_Delay_Occurrence_in_Transportation__1_.pdf |
_version_ |
1779157186583199744 |