Delayed Observation Planning in Partially Observable Domains
Traditional models for planning under uncertainty such as Markov Decision Processes (MDPs) or Partially Observable MDPs (POMDPs) assume that the observations about the results of agent actions are instantly available to the agent. In so doing, they are no longer applicable to domains where observati...
Saved in:
Main Authors: | , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2012
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/1606 https://ink.library.smu.edu.sg/context/sis_research/article/2605/type/native/viewcontent/citation.cfm_id_2343939 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-2605 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-26052012-11-14T04:06:31Z Delayed Observation Planning in Partially Observable Domains VARAKANTHAM, Pradeep Reddy Marecki, Janusz Traditional models for planning under uncertainty such as Markov Decision Processes (MDPs) or Partially Observable MDPs (POMDPs) assume that the observations about the results of agent actions are instantly available to the agent. In so doing, they are no longer applicable to domains where observations are received with delays caused by temporary unavailability of information (e.g. delayed response of the market to a new product). To that end, we make the following key contributions towards solving Delayed observation POMDPs (D-POMDPs): (i) We first provide an parameterized approximate algorithm for solving D-POMDPs efficiently, with desired accuracy; and (ii) We then propose a policy execution technique that adjusts the policy at run-time to account for the actual realization of observations. We then show the performance of our techniques on POMDP benchmark problems with delayed observations where explicit modeling of delayed observations leads to solutions of superior quality. 2012-06-01T07:00:00Z text text/html https://ink.library.smu.edu.sg/sis_research/1606 https://ink.library.smu.edu.sg/context/sis_research/article/2605/type/native/viewcontent/citation.cfm_id_2343939 Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Partially Observable Markov Decision Process Delayed Observations Artificial Intelligence and Robotics Business Operations Research, Systems Engineering and Industrial Engineering |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Partially Observable Markov Decision Process Delayed Observations Artificial Intelligence and Robotics Business Operations Research, Systems Engineering and Industrial Engineering |
spellingShingle |
Partially Observable Markov Decision Process Delayed Observations Artificial Intelligence and Robotics Business Operations Research, Systems Engineering and Industrial Engineering VARAKANTHAM, Pradeep Reddy Marecki, Janusz Delayed Observation Planning in Partially Observable Domains |
description |
Traditional models for planning under uncertainty such as Markov Decision Processes (MDPs) or Partially Observable MDPs (POMDPs) assume that the observations about the results of agent actions are instantly available to the agent. In so doing, they are no longer applicable to domains where observations are received with delays caused by temporary unavailability of information (e.g. delayed response of the market to a new product). To that end, we make the following key contributions towards solving Delayed observation POMDPs (D-POMDPs): (i) We first provide an parameterized approximate algorithm for solving D-POMDPs efficiently, with desired accuracy; and (ii) We then propose a policy execution technique that adjusts the policy at run-time to account for the actual realization of observations. We then show the performance of our techniques on POMDP benchmark problems with delayed observations where explicit modeling of delayed observations leads to solutions of superior quality. |
format |
text |
author |
VARAKANTHAM, Pradeep Reddy Marecki, Janusz |
author_facet |
VARAKANTHAM, Pradeep Reddy Marecki, Janusz |
author_sort |
VARAKANTHAM, Pradeep Reddy |
title |
Delayed Observation Planning in Partially Observable Domains |
title_short |
Delayed Observation Planning in Partially Observable Domains |
title_full |
Delayed Observation Planning in Partially Observable Domains |
title_fullStr |
Delayed Observation Planning in Partially Observable Domains |
title_full_unstemmed |
Delayed Observation Planning in Partially Observable Domains |
title_sort |
delayed observation planning in partially observable domains |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2012 |
url |
https://ink.library.smu.edu.sg/sis_research/1606 https://ink.library.smu.edu.sg/context/sis_research/article/2605/type/native/viewcontent/citation.cfm_id_2343939 |
_version_ |
1770571347651461120 |