Decentralized planning in stochastic environments with submodular rewards
Decentralized Markov Decision Process (Dec-MDP) providesa rich framework to represent cooperative decentralizedand stochastic planning problems under transition uncertainty.However, solving a Dec-MDP to generate coordinatedyet decentralized policies is NEXP-Hard. Researchershave made significant pro...
Saved in:
Main Authors: | , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2017
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/3549 https://ink.library.smu.edu.sg/context/sis_research/article/4550/viewcontent/14928_66557_1_PB.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-4550 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-45502019-06-25T13:05:02Z Decentralized planning in stochastic environments with submodular rewards KUMAR, Rajiv Ranjan Pradeep VARAKANTHAM, Akshat KUMAR, Decentralized Markov Decision Process (Dec-MDP) providesa rich framework to represent cooperative decentralizedand stochastic planning problems under transition uncertainty.However, solving a Dec-MDP to generate coordinatedyet decentralized policies is NEXP-Hard. Researchershave made significant progress in providing approximate approachesto improve scalability with respect to number ofagents. However, there has been little or no research devotedto finding guarantees on solution quality for approximateapproaches considering multiple (more than 2 agents)agents. We have a similar situation with respect to the competitivedecentralized planning problem and the StochasticGame (SG) model. To address this, we identify models in thecooperative and competitive case that rely on submodular rewards,where we show that existing approximate approachescan provide strong quality guarantees (a priori, and for cooperativecase also posteriori guarantees). We then providesolution approaches and demonstrate improved online guaranteeson benchmark problems from the literature for the cooperativecase. 2017-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/3549 https://ink.library.smu.edu.sg/context/sis_research/article/4550/viewcontent/14928_66557_1_PB.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Multiagent Systems Planning under uncertainty Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Multiagent Systems Planning under uncertainty Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering |
spellingShingle |
Multiagent Systems Planning under uncertainty Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering KUMAR, Rajiv Ranjan Pradeep VARAKANTHAM, Akshat KUMAR, Decentralized planning in stochastic environments with submodular rewards |
description |
Decentralized Markov Decision Process (Dec-MDP) providesa rich framework to represent cooperative decentralizedand stochastic planning problems under transition uncertainty.However, solving a Dec-MDP to generate coordinatedyet decentralized policies is NEXP-Hard. Researchershave made significant progress in providing approximate approachesto improve scalability with respect to number ofagents. However, there has been little or no research devotedto finding guarantees on solution quality for approximateapproaches considering multiple (more than 2 agents)agents. We have a similar situation with respect to the competitivedecentralized planning problem and the StochasticGame (SG) model. To address this, we identify models in thecooperative and competitive case that rely on submodular rewards,where we show that existing approximate approachescan provide strong quality guarantees (a priori, and for cooperativecase also posteriori guarantees). We then providesolution approaches and demonstrate improved online guaranteeson benchmark problems from the literature for the cooperativecase. |
format |
text |
author |
KUMAR, Rajiv Ranjan Pradeep VARAKANTHAM, Akshat KUMAR, |
author_facet |
KUMAR, Rajiv Ranjan Pradeep VARAKANTHAM, Akshat KUMAR, |
author_sort |
KUMAR, Rajiv Ranjan |
title |
Decentralized planning in stochastic environments with submodular rewards |
title_short |
Decentralized planning in stochastic environments with submodular rewards |
title_full |
Decentralized planning in stochastic environments with submodular rewards |
title_fullStr |
Decentralized planning in stochastic environments with submodular rewards |
title_full_unstemmed |
Decentralized planning in stochastic environments with submodular rewards |
title_sort |
decentralized planning in stochastic environments with submodular rewards |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2017 |
url |
https://ink.library.smu.edu.sg/sis_research/3549 https://ink.library.smu.edu.sg/context/sis_research/article/4550/viewcontent/14928_66557_1_PB.pdf |
_version_ |
1770573300770013184 |