Cooperative reinforcement learning in topology-based multi-agent systems
Topology-based multi-agent systems (TMAS), wherein agents interact with one another according to their spatial relationship in a network, are well suited for problems with topological constraints. In a TMAS system, however, each agent may have a different state space, which can be rather large. Cons...
Saved in:
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2014
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/101896 http://hdl.handle.net/10220/19819 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-101896 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1018962020-05-28T07:17:25Z Cooperative reinforcement learning in topology-based multi-agent systems Xiao, Dan Tan, Ah-Hwee School of Computer Engineering DRNTU::Engineering::Computer science and engineering Topology-based multi-agent systems (TMAS), wherein agents interact with one another according to their spatial relationship in a network, are well suited for problems with topological constraints. In a TMAS system, however, each agent may have a different state space, which can be rather large. Consequently, traditional approaches to multi-agent cooperative learning may not be able to scale up with the complexity of the network topology. In this paper, we propose a cooperative learning strategy, under which autonomous agents are assembled in a binary tree formation (BTF). By constraining the interaction between agents, we effectively unify the state space of individual agents and enable policy sharing across agents. Our complexity analysis indicates that multi-agent systems with the BTF have a much smaller state space and a higher level of flexibility, compared with the general form of n-ary (n > 2) tree formation. We have applied the proposed cooperative learning strategy to a class of reinforcement learning agents known as temporal difference-fusion architecture for learning and cognition (TD-FALCON). Comparative experiments based on a generic network routing problem, which is a typical TMAS domain, show that the TD-FALCON BTF teams outperform alternative methods, including TD-FALCON teams in single agent and n-ary tree formation, a Q-learning method based on the table lookup mechanism, as well as a classical linear programming algorithm. Our study further shows that TD-FALCON BTF can adapt and function well under various scales of network complexity and traffic volume in TMAS domains. 2014-06-19T02:35:16Z 2019-12-06T20:46:25Z 2014-06-19T02:35:16Z 2019-12-06T20:46:25Z 2011 2011 Journal Article Xiao, D., & Tan, A.-H. (2013). Cooperative reinforcement learning in topology-based multi-agent systems. Autonomous Agents and Multi-Agent Systems, 26(1), 86-119. 1387-2532 https://hdl.handle.net/10356/101896 http://hdl.handle.net/10220/19819 10.1007/s10458-011-9183-4 en Autonomous agents and multi-agent systems © 2011 The Author(s). |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering |
spellingShingle |
DRNTU::Engineering::Computer science and engineering Xiao, Dan Tan, Ah-Hwee Cooperative reinforcement learning in topology-based multi-agent systems |
description |
Topology-based multi-agent systems (TMAS), wherein agents interact with one another according to their spatial relationship in a network, are well suited for problems with topological constraints. In a TMAS system, however, each agent may have a different state space, which can be rather large. Consequently, traditional approaches to multi-agent cooperative learning may not be able to scale up with the complexity of the network topology. In this paper, we propose a cooperative learning strategy, under which autonomous agents are assembled in a binary tree formation (BTF). By constraining the interaction between agents, we effectively unify the state space of individual agents and enable policy sharing across agents. Our complexity analysis indicates that multi-agent systems with the BTF have a much smaller state space and a higher level of flexibility, compared with the general form of n-ary (n > 2) tree formation. We have applied the proposed cooperative learning strategy to a class of reinforcement learning agents known as temporal difference-fusion architecture for learning and cognition (TD-FALCON). Comparative experiments based on a generic network routing problem, which is a typical TMAS domain, show that the TD-FALCON BTF teams outperform alternative methods, including TD-FALCON teams in single agent and n-ary tree formation, a Q-learning method based on the table lookup mechanism, as well as a classical linear programming algorithm. Our study further shows that TD-FALCON BTF can adapt and function well under various scales of network complexity and traffic volume in TMAS domains. |
author2 |
School of Computer Engineering |
author_facet |
School of Computer Engineering Xiao, Dan Tan, Ah-Hwee |
format |
Article |
author |
Xiao, Dan Tan, Ah-Hwee |
author_sort |
Xiao, Dan |
title |
Cooperative reinforcement learning in topology-based multi-agent systems |
title_short |
Cooperative reinforcement learning in topology-based multi-agent systems |
title_full |
Cooperative reinforcement learning in topology-based multi-agent systems |
title_fullStr |
Cooperative reinforcement learning in topology-based multi-agent systems |
title_full_unstemmed |
Cooperative reinforcement learning in topology-based multi-agent systems |
title_sort |
cooperative reinforcement learning in topology-based multi-agent systems |
publishDate |
2014 |
url |
https://hdl.handle.net/10356/101896 http://hdl.handle.net/10220/19819 |
_version_ |
1681056072236793856 |