Optimal persistent monitoring using reinforcement learning

When monitoring a dynamically changing environment where a stationary group of agents cannot fully cover, a persistent monitoring problem (PMP) arises. In contrast to constantly monitoring, where every target must be monitored simultaneously, persistent monitoring requires a smaller number of agents...

Full description

Saved in:
Bibliographic Details
Main Author: Hu, Litao
Other Authors: Wen Changyun
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/149369
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-149369
record_format dspace
spelling sg-ntu-dr.10356-1493692023-07-07T18:13:29Z Optimal persistent monitoring using reinforcement learning Hu, Litao Wen Changyun School of Electrical and Electronic Engineering Liu Xiaokang ECYWEN@ntu.edu.sg Engineering::Electrical and electronic engineering When monitoring a dynamically changing environment where a stationary group of agents cannot fully cover, a persistent monitoring problem (PMP) arises. In contrast to constantly monitoring, where every target must be monitored simultaneously, persistent monitoring requires a smaller number of agents and provides an effective and reliable prediction with a minimized uncertainty metric. This project aims to implement Reinforcement Learning (RL) in the multiple targets monitoring simulation with a single agent. This paper presents a comparative analysis of five implementations in Reinforcement Learning: Deep Q Network (DQN), Double Deep Q Network (DDQN), Dueling Deep Q Network (Dueling DQN), Multi-Objective Deep Reinforcement Learning (MODRL) and Hierarchical Deep Q Network (HDQN). Different designs of the reward function and stop condition are tested and evaluated to improve models’ decision capability. This paper presents experiences in applying the goal decomposition, a new approach to feature extension to solve the persistent monitoring problem without modifying images, and an improved method for a highly dynamic environment. These proposed approaches significantly enhance the model’s performance and stability. Bachelor of Engineering (Electrical and Electronic Engineering) 2021-05-31T00:00:10Z 2021-05-31T00:00:10Z 2021 Final Year Project (FYP) Hu, L. (2021). Optimal persistent monitoring using reinforcement learning. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/149369 https://hdl.handle.net/10356/149369 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
spellingShingle Engineering::Electrical and electronic engineering
Hu, Litao
Optimal persistent monitoring using reinforcement learning
description When monitoring a dynamically changing environment where a stationary group of agents cannot fully cover, a persistent monitoring problem (PMP) arises. In contrast to constantly monitoring, where every target must be monitored simultaneously, persistent monitoring requires a smaller number of agents and provides an effective and reliable prediction with a minimized uncertainty metric. This project aims to implement Reinforcement Learning (RL) in the multiple targets monitoring simulation with a single agent. This paper presents a comparative analysis of five implementations in Reinforcement Learning: Deep Q Network (DQN), Double Deep Q Network (DDQN), Dueling Deep Q Network (Dueling DQN), Multi-Objective Deep Reinforcement Learning (MODRL) and Hierarchical Deep Q Network (HDQN). Different designs of the reward function and stop condition are tested and evaluated to improve models’ decision capability. This paper presents experiences in applying the goal decomposition, a new approach to feature extension to solve the persistent monitoring problem without modifying images, and an improved method for a highly dynamic environment. These proposed approaches significantly enhance the model’s performance and stability.
author2 Wen Changyun
author_facet Wen Changyun
Hu, Litao
format Final Year Project
author Hu, Litao
author_sort Hu, Litao
title Optimal persistent monitoring using reinforcement learning
title_short Optimal persistent monitoring using reinforcement learning
title_full Optimal persistent monitoring using reinforcement learning
title_fullStr Optimal persistent monitoring using reinforcement learning
title_full_unstemmed Optimal persistent monitoring using reinforcement learning
title_sort optimal persistent monitoring using reinforcement learning
publisher Nanyang Technological University
publishDate 2021
url https://hdl.handle.net/10356/149369
_version_ 1772828744358035456