Hierarchical multi-agent reinforcement learning with options
In recent years, there are many state-of-the-art multi-agent reinforcement learning (MARL) algorithms that aim to get multiple agents to work together to achieve a common goal. COMA is one of these breakthroughs that proposes a counterfactual baseline to address the credit assignment problem in mult...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/148028 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-148028 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1480282021-04-22T05:39:31Z Hierarchical multi-agent reinforcement learning with options Ang, Wan Qi Lana Obraztsova School of Computer Science and Engineering lana@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence In recent years, there are many state-of-the-art multi-agent reinforcement learning (MARL) algorithms that aim to get multiple agents to work together to achieve a common goal. COMA is one of these breakthroughs that proposes a counterfactual baseline to address the credit assignment problem in multi-agent scenarios. On the other hand, another research area known as hierarchical reinforcement learning (HRL) is gaining traction for its ability to model actions as a set of sub-actions to achieve temporal abstraction such as the options framework. This project aims to incorporate hierarchical reinforcement learning (HRL) frameworks into current state-of-the-art multi-agent reinforcement learning (MARL) algorithms through the use of the well-known options framework. As such, a novel MARL algorithm, COMA-OC, is introduced and evaluated on the SMAC benchmark in a variety of scenarios to analyse its performance and applicability. The results showed that COMA-OC is able to surpass the baseline's performance in most scenarios and has the potential to be developed further. Bachelor of Engineering Science (Computer Science) 2021-04-22T05:39:30Z 2021-04-22T05:39:30Z 2021 Final Year Project (FYP) Ang, W. Q. (2021). Hierarchical multi-agent reinforcement learning with options. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/148028 https://hdl.handle.net/10356/148028 en SCSE20-0490 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Ang, Wan Qi Hierarchical multi-agent reinforcement learning with options |
description |
In recent years, there are many state-of-the-art multi-agent reinforcement learning (MARL) algorithms that aim to get multiple agents to work together to achieve a common goal. COMA is one of these breakthroughs that proposes a counterfactual baseline to address the credit assignment problem in multi-agent scenarios. On the other hand, another research area known as hierarchical reinforcement learning (HRL) is gaining traction for its ability to model actions as a set of sub-actions to achieve temporal abstraction such as the options framework. This project aims to incorporate hierarchical reinforcement learning (HRL) frameworks into current state-of-the-art multi-agent reinforcement learning (MARL) algorithms through the use of the well-known options framework. As such, a novel MARL algorithm, COMA-OC, is introduced and evaluated on the SMAC benchmark in a variety of scenarios to analyse its performance and applicability. The results showed that COMA-OC is able to surpass the baseline's performance in most scenarios and has the potential to be developed further. |
author2 |
Lana Obraztsova |
author_facet |
Lana Obraztsova Ang, Wan Qi |
format |
Final Year Project |
author |
Ang, Wan Qi |
author_sort |
Ang, Wan Qi |
title |
Hierarchical multi-agent reinforcement learning with options |
title_short |
Hierarchical multi-agent reinforcement learning with options |
title_full |
Hierarchical multi-agent reinforcement learning with options |
title_fullStr |
Hierarchical multi-agent reinforcement learning with options |
title_full_unstemmed |
Hierarchical multi-agent reinforcement learning with options |
title_sort |
hierarchical multi-agent reinforcement learning with options |
publisher |
Nanyang Technological University |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/148028 |
_version_ |
1698713706302537728 |