Sampled fictitious play for multi-action stochastic dynamic programs

We introduce a class of finite-horizon dynamic optimization problems that we call multi-action stochastic dynamic programs (DPs). Their distinguishing feature is that the decision in each state is a multi-dimensional vector. These problems can in principle be solved using Bellman's backward rec...

Full description

Saved in:
Bibliographic Details
Main Authors: Ghate, Archis, CHENG, Shih-Fen, Baumert, Stephen, Reaume, Daniel, Sharma, Dushyant, Smith, Robert L.
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2014
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/1982
https://ink.library.smu.edu.sg/context/sis_research/article/2981/viewcontent/multi_action_stochastic_DP_final_production.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-2981
record_format dspace
spelling sg-smu-ink.sis_research-29812020-01-24T00:44:19Z Sampled fictitious play for multi-action stochastic dynamic programs Ghate, Archis CHENG, Shih-Fen Baumert, Stephen Reaume, Daniel Sharma, Dushyant Smith, Robert L. We introduce a class of finite-horizon dynamic optimization problems that we call multi-action stochastic dynamic programs (DPs). Their distinguishing feature is that the decision in each state is a multi-dimensional vector. These problems can in principle be solved using Bellman's backward recursion. However, complexity of this procedure grows exponentially in the dimension of the decision vectors. This is called the curse of action-space dimensionality. To overcome this computational challenge, we propose an approximation algorithm rooted in the game theoretic paradigm of Sampled Fictitious Play (SFP). SFP solves a sequence of DPs with a one-dimensional action-space, which are exponentially smaller than the original multi-action stochastic DP. In particular, the computational effort in a fixed number of SFP iterations is linear in the dimension of the decision vectors. We show that the sequence of SFP iterates converges to a local optimum, and present a numerical case study in manufacturing where SFP is able to find solutions with objective values within 1% of the optimal objective value hundreds of times faster than the time taken by backward recursion. In this case study, SFP solutions are also better by a statistically significant margin than those found by a one-step lookahead heuristic. 2014-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/1982 info:doi/10.1080/0740817X.2013.857062 https://ink.library.smu.edu.sg/context/sis_research/article/2981/viewcontent/multi_action_stochastic_DP_final_production.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Approximate dynamic programming game theory operations management Computer Sciences Operations Research, Systems Engineering and Industrial Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Approximate dynamic programming
game theory
operations management
Computer Sciences
Operations Research, Systems Engineering and Industrial Engineering
spellingShingle Approximate dynamic programming
game theory
operations management
Computer Sciences
Operations Research, Systems Engineering and Industrial Engineering
Ghate, Archis
CHENG, Shih-Fen
Baumert, Stephen
Reaume, Daniel
Sharma, Dushyant
Smith, Robert L.
Sampled fictitious play for multi-action stochastic dynamic programs
description We introduce a class of finite-horizon dynamic optimization problems that we call multi-action stochastic dynamic programs (DPs). Their distinguishing feature is that the decision in each state is a multi-dimensional vector. These problems can in principle be solved using Bellman's backward recursion. However, complexity of this procedure grows exponentially in the dimension of the decision vectors. This is called the curse of action-space dimensionality. To overcome this computational challenge, we propose an approximation algorithm rooted in the game theoretic paradigm of Sampled Fictitious Play (SFP). SFP solves a sequence of DPs with a one-dimensional action-space, which are exponentially smaller than the original multi-action stochastic DP. In particular, the computational effort in a fixed number of SFP iterations is linear in the dimension of the decision vectors. We show that the sequence of SFP iterates converges to a local optimum, and present a numerical case study in manufacturing where SFP is able to find solutions with objective values within 1% of the optimal objective value hundreds of times faster than the time taken by backward recursion. In this case study, SFP solutions are also better by a statistically significant margin than those found by a one-step lookahead heuristic.
format text
author Ghate, Archis
CHENG, Shih-Fen
Baumert, Stephen
Reaume, Daniel
Sharma, Dushyant
Smith, Robert L.
author_facet Ghate, Archis
CHENG, Shih-Fen
Baumert, Stephen
Reaume, Daniel
Sharma, Dushyant
Smith, Robert L.
author_sort Ghate, Archis
title Sampled fictitious play for multi-action stochastic dynamic programs
title_short Sampled fictitious play for multi-action stochastic dynamic programs
title_full Sampled fictitious play for multi-action stochastic dynamic programs
title_fullStr Sampled fictitious play for multi-action stochastic dynamic programs
title_full_unstemmed Sampled fictitious play for multi-action stochastic dynamic programs
title_sort sampled fictitious play for multi-action stochastic dynamic programs
publisher Institutional Knowledge at Singapore Management University
publishDate 2014
url https://ink.library.smu.edu.sg/sis_research/1982
https://ink.library.smu.edu.sg/context/sis_research/article/2981/viewcontent/multi_action_stochastic_DP_final_production.pdf
_version_ 1770571763553402880