A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative traject...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2025
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/182740 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-182740 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1827402025-02-25T15:31:40Z A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning Pang, Bizhao Hu, Xinting Zhang, Mingcheng Alam, Sameer Lulli, Guglielmo School of Mechanical and Aerospace Engineering Air Traffic Management Research Institute Engineering Air traffic management Autonomous trajectory planning Deep reinforcement learning Dynamic thunderstorm cells Climate change Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single-agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths. ETA uncertainty analysis demonstrated the model's robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations. Agency for Science, Technology and Research (A*STAR) Submitted/Accepted version This research is supported by the Italian Ministry of Foreign Affairs and International Cooperation (MAECI) and the Agency for Science, Technology and Research (A*STAR), Singapore, under the First Executive Programme of Scientific and Technological Cooperation between Italy and Singapore for the years 2023–2025. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the Italian Ministry of Foreign Affairs and International Cooperation or the Agency for Science, Technology and Research (A*STAR), Singapore. 2025-02-20T11:13:36Z 2025-02-20T11:13:36Z 2025 Journal Article Pang, B., Hu, X., Zhang, M., Alam, S. & Lulli, G. (2025). A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning. Advanced Engineering Informatics, 65, 103157-. https://dx.doi.org/10.1016/j.aei.2025.103157 1474-0346 https://hdl.handle.net/10356/182740 10.1016/j.aei.2025.103157 2-s2.0-85216669279 65 103157 en Advanced Engineering Informatics © 2025 Elsevier Ltd.. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1016/j.aei.2025.103157. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Air traffic management Autonomous trajectory planning Deep reinforcement learning Dynamic thunderstorm cells Climate change |
spellingShingle |
Engineering Air traffic management Autonomous trajectory planning Deep reinforcement learning Dynamic thunderstorm cells Climate change Pang, Bizhao Hu, Xinting Zhang, Mingcheng Alam, Sameer Lulli, Guglielmo A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
description |
Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single-agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths. ETA uncertainty analysis demonstrated the model's robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations. |
author2 |
School of Mechanical and Aerospace Engineering |
author_facet |
School of Mechanical and Aerospace Engineering Pang, Bizhao Hu, Xinting Zhang, Mingcheng Alam, Sameer Lulli, Guglielmo |
format |
Article |
author |
Pang, Bizhao Hu, Xinting Zhang, Mingcheng Alam, Sameer Lulli, Guglielmo |
author_sort |
Pang, Bizhao |
title |
A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
title_short |
A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
title_full |
A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
title_fullStr |
A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
title_full_unstemmed |
A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
title_sort |
multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/182740 |
_version_ |
1825619638184050688 |