Collaborative air-ground search with deep reinforcement learning
Artificial intelligence (AI) has emerged as a leading area of research, particularly in the realm of training autonomous Unmanned Aerial Vehicles (UAVs). Target searching, a key focus within this domain, holds significant promise for applications such as runway approach, cargo pickup and delivery, a...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177158 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-177158 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1771582024-05-25T16:50:25Z Collaborative air-ground search with deep reinforcement learning Lim, You Xuan Mir Feroskhan School of Mechanical and Aerospace Engineering mir.feroskhan@ntu.edu.sg Computer and Information Science Engineering Deep reinforcement learning Intelligent agents Collaborative search Artificial intelligence (AI) has emerged as a leading area of research, particularly in the realm of training autonomous Unmanned Aerial Vehicles (UAVs). Target searching, a key focus within this domain, holds significant promise for applications such as runway approach, cargo pickup and delivery, and area surveillance. However, current target searching methods entail intensive path planning efforts to ensure precise drone actions. To address these complexities, reinforcement learning algorithms like Proximal Policy Optimization (PPO) have been employed to train autonomous behaviors. While PPO has shown promise, it lacks innate support for collaborative behaviors crucial to the success of drone swarms. In response to these challenges, this study extends beyond UAV training to incorporate Unmanned Ground Vehicles (UGVs), recognizing the limitations faced by each platform individually. Navigating environments solely with UGVs is hindered by limited observational capabilities, while relying solely on UAVs presents challenges in carrying heavy loads during search and rescue missions. To address these limitations, a dual-agent approach is adopted, training UAVs to locate targets and plan paths while leading UGVs. This integrated strategy effectively addresses navigational and load-bearing challenges, optimizing performance in real-world scenarios. In this study, we propose a Cooperative Multi-goal Multi-stage Multi-agent (CM3) Deep Reinforcement Learning approach for target search missions, aiming to address the inherent complexities and challenges in coordinating UAVs and UGVs effectively. By incorporating the CM3 framework, our approach facilitates seamless collaboration between multiple agents across different stages of the mission, thereby enhancing overall performance and efficiency. Training strategies using PPO and the Multi-Agent Posthumous Credit Assignment (MA-POCA) algorithm are also investigated to foster collaborative behaviors within the drone swarm. Notably, the study reveals insights into the impact of camera resolution on target search effectiveness and the importance of careful consideration when implementing group reward mechanisms. Additionally, the study highlights the necessity of aligning group reward assignments with agents' training stages in multi-stage reinforcement learning scenarios. In summary, this research underscores the significance of AI in enhancing UAV and UGV capabilities for target searching missions. By leveraging reinforcement learning algorithms and adopting a collaborative approach, the study offers valuable insights into optimizing performance and addressing challenges in real-world applications. Bachelor's degree 2024-05-21T09:26:05Z 2024-05-21T09:26:05Z 2024 Final Year Project (FYP) Lim, Y. X. (2024). Collaborative air-ground search with deep reinforcement learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177158 https://hdl.handle.net/10356/177158 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Engineering Deep reinforcement learning Intelligent agents Collaborative search |
spellingShingle |
Computer and Information Science Engineering Deep reinforcement learning Intelligent agents Collaborative search Lim, You Xuan Collaborative air-ground search with deep reinforcement learning |
description |
Artificial intelligence (AI) has emerged as a leading area of research, particularly in the realm of training autonomous Unmanned Aerial Vehicles (UAVs). Target searching, a key focus within this domain, holds significant promise for applications such as runway approach, cargo pickup and delivery, and area surveillance. However, current target searching methods entail intensive path planning efforts to ensure precise drone actions. To address these complexities, reinforcement learning algorithms like Proximal Policy Optimization (PPO) have been employed to train autonomous behaviors. While PPO has shown promise, it lacks innate support for collaborative behaviors crucial to the success of drone swarms.
In response to these challenges, this study extends beyond UAV training to incorporate Unmanned Ground Vehicles (UGVs), recognizing the limitations faced by each platform individually. Navigating environments solely with UGVs is hindered by limited observational capabilities, while relying solely on UAVs presents challenges in carrying heavy loads during search and rescue missions. To address these limitations, a dual-agent approach is adopted, training UAVs to locate targets and plan paths while leading UGVs. This integrated strategy effectively addresses navigational and load-bearing challenges, optimizing performance in real-world scenarios.
In this study, we propose a Cooperative Multi-goal Multi-stage Multi-agent (CM3) Deep Reinforcement Learning approach for target search missions, aiming to address the inherent complexities and challenges in coordinating UAVs and UGVs effectively. By incorporating the CM3 framework, our approach facilitates seamless collaboration between multiple agents across different stages of the mission, thereby enhancing overall performance and efficiency.
Training strategies using PPO and the Multi-Agent Posthumous Credit Assignment (MA-POCA) algorithm are also investigated to foster collaborative behaviors within the drone swarm. Notably, the study reveals insights into the impact of camera resolution on target search effectiveness and the importance of careful consideration when implementing group reward mechanisms. Additionally, the study highlights the necessity of aligning group reward assignments with agents' training stages in multi-stage reinforcement learning scenarios.
In summary, this research underscores the significance of AI in enhancing UAV and UGV capabilities for target searching missions. By leveraging reinforcement learning algorithms and adopting a collaborative approach, the study offers valuable insights into optimizing performance and addressing challenges in real-world applications. |
author2 |
Mir Feroskhan |
author_facet |
Mir Feroskhan Lim, You Xuan |
format |
Final Year Project |
author |
Lim, You Xuan |
author_sort |
Lim, You Xuan |
title |
Collaborative air-ground search with deep reinforcement learning |
title_short |
Collaborative air-ground search with deep reinforcement learning |
title_full |
Collaborative air-ground search with deep reinforcement learning |
title_fullStr |
Collaborative air-ground search with deep reinforcement learning |
title_full_unstemmed |
Collaborative air-ground search with deep reinforcement learning |
title_sort |
collaborative air-ground search with deep reinforcement learning |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/177158 |
_version_ |
1800916147496812544 |