A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning

Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative traject...

Full description

Saved in:
Bibliographic Details
Main Authors: Pang, Bizhao, Hu, Xinting, Zhang, Mingcheng, Alam, Sameer, Lulli, Guglielmo
Other Authors: School of Mechanical and Aerospace Engineering
Format: Article
Language:English
Published: 2025
Subjects:
Online Access:https://hdl.handle.net/10356/182740
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-182740
record_format dspace
spelling sg-ntu-dr.10356-1827402025-02-25T15:31:40Z A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning Pang, Bizhao Hu, Xinting Zhang, Mingcheng Alam, Sameer Lulli, Guglielmo School of Mechanical and Aerospace Engineering Air Traffic Management Research Institute Engineering Air traffic management Autonomous trajectory planning Deep reinforcement learning Dynamic thunderstorm cells Climate change Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single-agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths. ETA uncertainty analysis demonstrated the model's robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations. Agency for Science, Technology and Research (A*STAR) Submitted/Accepted version This research is supported by the Italian Ministry of Foreign Affairs and International Cooperation (MAECI) and the Agency for Science, Technology and Research (A*STAR), Singapore, under the First Executive Programme of Scientific and Technological Cooperation between Italy and Singapore for the years 2023–2025. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the Italian Ministry of Foreign Affairs and International Cooperation or the Agency for Science, Technology and Research (A*STAR), Singapore. 2025-02-20T11:13:36Z 2025-02-20T11:13:36Z 2025 Journal Article Pang, B., Hu, X., Zhang, M., Alam, S. & Lulli, G. (2025). A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning. Advanced Engineering Informatics, 65, 103157-. https://dx.doi.org/10.1016/j.aei.2025.103157 1474-0346 https://hdl.handle.net/10356/182740 10.1016/j.aei.2025.103157 2-s2.0-85216669279 65 103157 en Advanced Engineering Informatics © 2025 Elsevier Ltd.. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1016/j.aei.2025.103157. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Air traffic management
Autonomous trajectory planning
Deep reinforcement learning
Dynamic thunderstorm cells
Climate change
spellingShingle Engineering
Air traffic management
Autonomous trajectory planning
Deep reinforcement learning
Dynamic thunderstorm cells
Climate change
Pang, Bizhao
Hu, Xinting
Zhang, Mingcheng
Alam, Sameer
Lulli, Guglielmo
A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
description Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single-agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths. ETA uncertainty analysis demonstrated the model's robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations.
author2 School of Mechanical and Aerospace Engineering
author_facet School of Mechanical and Aerospace Engineering
Pang, Bizhao
Hu, Xinting
Zhang, Mingcheng
Alam, Sameer
Lulli, Guglielmo
format Article
author Pang, Bizhao
Hu, Xinting
Zhang, Mingcheng
Alam, Sameer
Lulli, Guglielmo
author_sort Pang, Bizhao
title A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
title_short A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
title_full A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
title_fullStr A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
title_full_unstemmed A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
title_sort multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning
publishDate 2025
url https://hdl.handle.net/10356/182740
_version_ 1825619638184050688