Cooperative reinforcement learning in topology-based multi-agent systems

Topology-based multi-agent systems (TMAS), wherein agents interact with one another according to their spatial relationship in a network, are well suited for problems with topological constraints. In a TMAS system, however, each agent may have a different state space, which can be rather large. Cons...

Full description

Saved in:
Bibliographic Details
Main Authors: Xiao, Dan, Tan, Ah-Hwee
Other Authors: School of Computer Engineering
Format: Article
Language:English
Published: 2014
Subjects:
Online Access:https://hdl.handle.net/10356/101896
http://hdl.handle.net/10220/19819
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-101896
record_format dspace
spelling sg-ntu-dr.10356-1018962020-05-28T07:17:25Z Cooperative reinforcement learning in topology-based multi-agent systems Xiao, Dan Tan, Ah-Hwee School of Computer Engineering DRNTU::Engineering::Computer science and engineering Topology-based multi-agent systems (TMAS), wherein agents interact with one another according to their spatial relationship in a network, are well suited for problems with topological constraints. In a TMAS system, however, each agent may have a different state space, which can be rather large. Consequently, traditional approaches to multi-agent cooperative learning may not be able to scale up with the complexity of the network topology. In this paper, we propose a cooperative learning strategy, under which autonomous agents are assembled in a binary tree formation (BTF). By constraining the interaction between agents, we effectively unify the state space of individual agents and enable policy sharing across agents. Our complexity analysis indicates that multi-agent systems with the BTF have a much smaller state space and a higher level of flexibility, compared with the general form of n-ary (n > 2) tree formation. We have applied the proposed cooperative learning strategy to a class of reinforcement learning agents known as temporal difference-fusion architecture for learning and cognition (TD-FALCON). Comparative experiments based on a generic network routing problem, which is a typical TMAS domain, show that the TD-FALCON BTF teams outperform alternative methods, including TD-FALCON teams in single agent and n-ary tree formation, a Q-learning method based on the table lookup mechanism, as well as a classical linear programming algorithm. Our study further shows that TD-FALCON BTF can adapt and function well under various scales of network complexity and traffic volume in TMAS domains. 2014-06-19T02:35:16Z 2019-12-06T20:46:25Z 2014-06-19T02:35:16Z 2019-12-06T20:46:25Z 2011 2011 Journal Article Xiao, D., & Tan, A.-H. (2013). Cooperative reinforcement learning in topology-based multi-agent systems. Autonomous Agents and Multi-Agent Systems, 26(1), 86-119. 1387-2532 https://hdl.handle.net/10356/101896 http://hdl.handle.net/10220/19819 10.1007/s10458-011-9183-4 en Autonomous agents and multi-agent systems © 2011 The Author(s).
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering
spellingShingle DRNTU::Engineering::Computer science and engineering
Xiao, Dan
Tan, Ah-Hwee
Cooperative reinforcement learning in topology-based multi-agent systems
description Topology-based multi-agent systems (TMAS), wherein agents interact with one another according to their spatial relationship in a network, are well suited for problems with topological constraints. In a TMAS system, however, each agent may have a different state space, which can be rather large. Consequently, traditional approaches to multi-agent cooperative learning may not be able to scale up with the complexity of the network topology. In this paper, we propose a cooperative learning strategy, under which autonomous agents are assembled in a binary tree formation (BTF). By constraining the interaction between agents, we effectively unify the state space of individual agents and enable policy sharing across agents. Our complexity analysis indicates that multi-agent systems with the BTF have a much smaller state space and a higher level of flexibility, compared with the general form of n-ary (n > 2) tree formation. We have applied the proposed cooperative learning strategy to a class of reinforcement learning agents known as temporal difference-fusion architecture for learning and cognition (TD-FALCON). Comparative experiments based on a generic network routing problem, which is a typical TMAS domain, show that the TD-FALCON BTF teams outperform alternative methods, including TD-FALCON teams in single agent and n-ary tree formation, a Q-learning method based on the table lookup mechanism, as well as a classical linear programming algorithm. Our study further shows that TD-FALCON BTF can adapt and function well under various scales of network complexity and traffic volume in TMAS domains.
author2 School of Computer Engineering
author_facet School of Computer Engineering
Xiao, Dan
Tan, Ah-Hwee
format Article
author Xiao, Dan
Tan, Ah-Hwee
author_sort Xiao, Dan
title Cooperative reinforcement learning in topology-based multi-agent systems
title_short Cooperative reinforcement learning in topology-based multi-agent systems
title_full Cooperative reinforcement learning in topology-based multi-agent systems
title_fullStr Cooperative reinforcement learning in topology-based multi-agent systems
title_full_unstemmed Cooperative reinforcement learning in topology-based multi-agent systems
title_sort cooperative reinforcement learning in topology-based multi-agent systems
publishDate 2014
url https://hdl.handle.net/10356/101896
http://hdl.handle.net/10220/19819
_version_ 1681056072236793856