Decentralized planning in stochastic environments with submodular rewards

Decentralized Markov Decision Process (Dec-MDP) providesa rich framework to represent cooperative decentralizedand stochastic planning problems under transition uncertainty.However, solving a Dec-MDP to generate coordinatedyet decentralized policies is NEXP-Hard. Researchershave made significant pro...

Full description

Saved in:
Bibliographic Details
Main Authors: KUMAR, Rajiv Ranjan, Pradeep VARAKANTHAM, Akshat KUMAR
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2017
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/3549
https://ink.library.smu.edu.sg/context/sis_research/article/4550/viewcontent/14928_66557_1_PB.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-4550
record_format dspace
spelling sg-smu-ink.sis_research-45502019-06-25T13:05:02Z Decentralized planning in stochastic environments with submodular rewards KUMAR, Rajiv Ranjan Pradeep VARAKANTHAM, Akshat KUMAR, Decentralized Markov Decision Process (Dec-MDP) providesa rich framework to represent cooperative decentralizedand stochastic planning problems under transition uncertainty.However, solving a Dec-MDP to generate coordinatedyet decentralized policies is NEXP-Hard. Researchershave made significant progress in providing approximate approachesto improve scalability with respect to number ofagents. However, there has been little or no research devotedto finding guarantees on solution quality for approximateapproaches considering multiple (more than 2 agents)agents. We have a similar situation with respect to the competitivedecentralized planning problem and the StochasticGame (SG) model. To address this, we identify models in thecooperative and competitive case that rely on submodular rewards,where we show that existing approximate approachescan provide strong quality guarantees (a priori, and for cooperativecase also posteriori guarantees). We then providesolution approaches and demonstrate improved online guaranteeson benchmark problems from the literature for the cooperativecase. 2017-02-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/3549 https://ink.library.smu.edu.sg/context/sis_research/article/4550/viewcontent/14928_66557_1_PB.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Multiagent Systems Planning under uncertainty Artificial Intelligence and Robotics Operations Research, Systems Engineering and Industrial Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Multiagent Systems
Planning under uncertainty
Artificial Intelligence and Robotics
Operations Research, Systems Engineering and Industrial Engineering
spellingShingle Multiagent Systems
Planning under uncertainty
Artificial Intelligence and Robotics
Operations Research, Systems Engineering and Industrial Engineering
KUMAR, Rajiv Ranjan
Pradeep VARAKANTHAM,
Akshat KUMAR,
Decentralized planning in stochastic environments with submodular rewards
description Decentralized Markov Decision Process (Dec-MDP) providesa rich framework to represent cooperative decentralizedand stochastic planning problems under transition uncertainty.However, solving a Dec-MDP to generate coordinatedyet decentralized policies is NEXP-Hard. Researchershave made significant progress in providing approximate approachesto improve scalability with respect to number ofagents. However, there has been little or no research devotedto finding guarantees on solution quality for approximateapproaches considering multiple (more than 2 agents)agents. We have a similar situation with respect to the competitivedecentralized planning problem and the StochasticGame (SG) model. To address this, we identify models in thecooperative and competitive case that rely on submodular rewards,where we show that existing approximate approachescan provide strong quality guarantees (a priori, and for cooperativecase also posteriori guarantees). We then providesolution approaches and demonstrate improved online guaranteeson benchmark problems from the literature for the cooperativecase.
format text
author KUMAR, Rajiv Ranjan
Pradeep VARAKANTHAM,
Akshat KUMAR,
author_facet KUMAR, Rajiv Ranjan
Pradeep VARAKANTHAM,
Akshat KUMAR,
author_sort KUMAR, Rajiv Ranjan
title Decentralized planning in stochastic environments with submodular rewards
title_short Decentralized planning in stochastic environments with submodular rewards
title_full Decentralized planning in stochastic environments with submodular rewards
title_fullStr Decentralized planning in stochastic environments with submodular rewards
title_full_unstemmed Decentralized planning in stochastic environments with submodular rewards
title_sort decentralized planning in stochastic environments with submodular rewards
publisher Institutional Knowledge at Singapore Management University
publishDate 2017
url https://ink.library.smu.edu.sg/sis_research/3549
https://ink.library.smu.edu.sg/context/sis_research/article/4550/viewcontent/14928_66557_1_PB.pdf
_version_ 1770573300770013184