Optimal persistent monitoring using reinforcement learning
When monitoring a dynamically changing environment where a stationary group of agents cannot fully cover, a persistent monitoring problem (PMP) arises. In contrast to constantly monitoring, where every target must be monitored simultaneously, persistent monitoring requires a smaller number of agents...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/149369 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-149369 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1493692023-07-07T18:13:29Z Optimal persistent monitoring using reinforcement learning Hu, Litao Wen Changyun School of Electrical and Electronic Engineering Liu Xiaokang ECYWEN@ntu.edu.sg Engineering::Electrical and electronic engineering When monitoring a dynamically changing environment where a stationary group of agents cannot fully cover, a persistent monitoring problem (PMP) arises. In contrast to constantly monitoring, where every target must be monitored simultaneously, persistent monitoring requires a smaller number of agents and provides an effective and reliable prediction with a minimized uncertainty metric. This project aims to implement Reinforcement Learning (RL) in the multiple targets monitoring simulation with a single agent. This paper presents a comparative analysis of five implementations in Reinforcement Learning: Deep Q Network (DQN), Double Deep Q Network (DDQN), Dueling Deep Q Network (Dueling DQN), Multi-Objective Deep Reinforcement Learning (MODRL) and Hierarchical Deep Q Network (HDQN). Different designs of the reward function and stop condition are tested and evaluated to improve models’ decision capability. This paper presents experiences in applying the goal decomposition, a new approach to feature extension to solve the persistent monitoring problem without modifying images, and an improved method for a highly dynamic environment. These proposed approaches significantly enhance the model’s performance and stability. Bachelor of Engineering (Electrical and Electronic Engineering) 2021-05-31T00:00:10Z 2021-05-31T00:00:10Z 2021 Final Year Project (FYP) Hu, L. (2021). Optimal persistent monitoring using reinforcement learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/149369 https://hdl.handle.net/10356/149369 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering |
spellingShingle |
Engineering::Electrical and electronic engineering Hu, Litao Optimal persistent monitoring using reinforcement learning |
description |
When monitoring a dynamically changing environment where a stationary group of agents cannot fully cover, a persistent monitoring problem (PMP) arises. In contrast to constantly monitoring, where every target must be monitored simultaneously, persistent monitoring requires a smaller number of agents and provides an effective and reliable prediction with a minimized uncertainty metric. This project aims to implement Reinforcement Learning (RL) in the multiple targets monitoring simulation with a single agent. This paper presents a comparative analysis of five implementations in Reinforcement Learning: Deep Q Network (DQN), Double Deep Q Network (DDQN), Dueling Deep Q Network (Dueling DQN), Multi-Objective Deep Reinforcement Learning (MODRL) and Hierarchical Deep Q Network (HDQN). Different designs of the reward function and stop condition are tested and evaluated to improve models’ decision capability. This paper presents experiences in applying the goal decomposition, a new approach to feature extension to solve the persistent monitoring problem without modifying images, and an improved method for a highly dynamic environment. These proposed approaches significantly enhance the model’s performance and stability. |
author2 |
Wen Changyun |
author_facet |
Wen Changyun Hu, Litao |
format |
Final Year Project |
author |
Hu, Litao |
author_sort |
Hu, Litao |
title |
Optimal persistent monitoring using reinforcement learning |
title_short |
Optimal persistent monitoring using reinforcement learning |
title_full |
Optimal persistent monitoring using reinforcement learning |
title_fullStr |
Optimal persistent monitoring using reinforcement learning |
title_full_unstemmed |
Optimal persistent monitoring using reinforcement learning |
title_sort |
optimal persistent monitoring using reinforcement learning |
publisher |
Nanyang Technological University |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/149369 |
_version_ |
1772828744358035456 |