Learning index policies for restless bandits with application to maternal healthcare
In many community health settings, it is crucial to have a systematic monitoring and intervention process to ensure that the patients adhere to healthcare programs, such as periodic health checks or taking medications. When these interventions are expensive, they can be provided to only a fixed smal...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/6774 https://ink.library.smu.edu.sg/context/sis_research/article/7777/viewcontent/Learning_Index_Policies_for_Restless_Bandits_with_Application_to_Maternal_Healthcare.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-7777 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-77772022-01-27T10:05:56Z Learning index policies for restless bandits with application to maternal healthcare BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind In many community health settings, it is crucial to have a systematic monitoring and intervention process to ensure that the patients adhere to healthcare programs, such as periodic health checks or taking medications. When these interventions are expensive, they can be provided to only a fixed small fraction of the patients at any period of time. Hence, it is important to carefully choose the beneficiaries who should be provided with interventions and when. We model this scenario as a restless multi-armed bandit (RMAB) problem, where each beneficiary is assumed to transition from one state to another depending on the intervention provided to them. In practice, the transition probabilities are unknown a priori, and hence, we propose a mechanism for the problem of balancing the explore-exploit trade-off. Empirically, we find that our proposed mechanism outperforms the baseline intervention scheme maternal healthcare dataset. 2021-05-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6774 https://ink.library.smu.edu.sg/context/sis_research/article/7777/viewcontent/Learning_Index_Policies_for_Restless_Bandits_with_Application_to_Maternal_Healthcare.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University reinforcement learning multi-armed bandits unknown transition probabilities Artificial Intelligence and Robotics Databases and Information Systems |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
reinforcement learning multi-armed bandits unknown transition probabilities Artificial Intelligence and Robotics Databases and Information Systems |
spellingShingle |
reinforcement learning multi-armed bandits unknown transition probabilities Artificial Intelligence and Robotics Databases and Information Systems BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind Learning index policies for restless bandits with application to maternal healthcare |
description |
In many community health settings, it is crucial to have a systematic monitoring and intervention process to ensure that the patients adhere to healthcare programs, such as periodic health checks or taking medications. When these interventions are expensive, they can be provided to only a fixed small fraction of the patients at any period of time. Hence, it is important to carefully choose the beneficiaries who should be provided with interventions and when. We model this scenario as a restless multi-armed bandit (RMAB) problem, where each beneficiary is assumed to transition from one state to another depending on the intervention provided to them. In practice, the transition probabilities are unknown a priori, and hence, we propose a mechanism for the problem of balancing the explore-exploit trade-off. Empirically, we find that our proposed mechanism outperforms the baseline intervention scheme maternal healthcare dataset. |
format |
text |
author |
BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind |
author_facet |
BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind |
author_sort |
BISWAS, Arpita |
title |
Learning index policies for restless bandits with application to maternal healthcare |
title_short |
Learning index policies for restless bandits with application to maternal healthcare |
title_full |
Learning index policies for restless bandits with application to maternal healthcare |
title_fullStr |
Learning index policies for restless bandits with application to maternal healthcare |
title_full_unstemmed |
Learning index policies for restless bandits with application to maternal healthcare |
title_sort |
learning index policies for restless bandits with application to maternal healthcare |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2021 |
url |
https://ink.library.smu.edu.sg/sis_research/6774 https://ink.library.smu.edu.sg/context/sis_research/article/7777/viewcontent/Learning_Index_Policies_for_Restless_Bandits_with_Application_to_Maternal_Healthcare.pdf |
_version_ |
1770576062170791936 |