Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies

Distributed Partially Observable Markov Decision Problems (Distributed POMDPs) are a popular approach for modeling multi-agent systems acting in uncertain domains. Given the significant complexity of solving distributed POMDPs, particularly as we scale up the numbers of agents, one popular approach...

Full description

Saved in:
Bibliographic Details
Main Authors: VARAKANTHAM, Pradeep Reddy, Marecki, Janusz, Yokoo, Makoto, Tambe, Milind
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2007
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/947
http://portal.acm.org/citation.cfm?id=1329388
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-1946
record_format dspace
spelling sg-smu-ink.sis_research-19462010-12-15T08:06:06Z Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies VARAKANTHAM, Pradeep Reddy Marecki, Janusz Yokoo, Makoto Tambe, Milind Distributed Partially Observable Markov Decision Problems (Distributed POMDPs) are a popular approach for modeling multi-agent systems acting in uncertain domains. Given the significant complexity of solving distributed POMDPs, particularly as we scale up the numbers of agents, one popular approach has focused on approximate solutions. Though this approach is efficient, the algorithms within this approach do not provide any guarantees on solution quality. A second less popular approach focuses on global optimality, but typical results are available only for two agents, and also at considerable computational cost. This paper overcomes the limitations of both these approaches by providing SPIDER, a novel combination of three key features for policy generation in distributed POMDPs: (i) it exploits agent interaction structure given a network of agents (i.e. allowing easier scale-up to larger number of agents); (ii) it uses a combination of heuristics to speedup policy search; and (iii) it allows quality guaranteed approximations, allowing a systematic tradeoff of solution quality for time. Experimental results show orders of magnitude improvement in performance when compared with previous global optimal algorithms. 2007-01-01T08:00:00Z text https://ink.library.smu.edu.sg/sis_research/947 info:doi/10.1145/1329125.1329388 http://portal.acm.org/citation.cfm?id=1329388 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Business Operations Research, Systems Engineering and Industrial Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence and Robotics
Business
Operations Research, Systems Engineering and Industrial Engineering
spellingShingle Artificial Intelligence and Robotics
Business
Operations Research, Systems Engineering and Industrial Engineering
VARAKANTHAM, Pradeep Reddy
Marecki, Janusz
Yokoo, Makoto
Tambe, Milind
Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies
description Distributed Partially Observable Markov Decision Problems (Distributed POMDPs) are a popular approach for modeling multi-agent systems acting in uncertain domains. Given the significant complexity of solving distributed POMDPs, particularly as we scale up the numbers of agents, one popular approach has focused on approximate solutions. Though this approach is efficient, the algorithms within this approach do not provide any guarantees on solution quality. A second less popular approach focuses on global optimality, but typical results are available only for two agents, and also at considerable computational cost. This paper overcomes the limitations of both these approaches by providing SPIDER, a novel combination of three key features for policy generation in distributed POMDPs: (i) it exploits agent interaction structure given a network of agents (i.e. allowing easier scale-up to larger number of agents); (ii) it uses a combination of heuristics to speedup policy search; and (iii) it allows quality guaranteed approximations, allowing a systematic tradeoff of solution quality for time. Experimental results show orders of magnitude improvement in performance when compared with previous global optimal algorithms.
format text
author VARAKANTHAM, Pradeep Reddy
Marecki, Janusz
Yokoo, Makoto
Tambe, Milind
author_facet VARAKANTHAM, Pradeep Reddy
Marecki, Janusz
Yokoo, Makoto
Tambe, Milind
author_sort VARAKANTHAM, Pradeep Reddy
title Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies
title_short Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies
title_full Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies
title_fullStr Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies
title_full_unstemmed Letting loose a SPIDER on a network of POMDPs: Generating quality guranteed policies
title_sort letting loose a spider on a network of pomdps: generating quality guranteed policies
publisher Institutional Knowledge at Singapore Management University
publishDate 2007
url https://ink.library.smu.edu.sg/sis_research/947
http://portal.acm.org/citation.cfm?id=1329388
_version_ 1770570780027912192