Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
In many public health settings, it is important for patients to adhere to health programs, such as taking medications and periodic health checks. Unfortunately, beneficiaries may gradually disengage from such programs, which is detrimental to their health. A concrete example of gradual disengagement...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/6794 https://ink.library.smu.edu.sg/context/sis_research/article/7797/viewcontent/0556.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-7797 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-77972022-01-27T09:57:40Z Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind In many public health settings, it is important for patients to adhere to health programs, such as taking medications and periodic health checks. Unfortunately, beneficiaries may gradually disengage from such programs, which is detrimental to their health. A concrete example of gradual disengagement has been observed by an organization that carries out a free automated call-based program for spreading preventive care information among pregnant women. Many women stop picking up calls after being enrolled for a few months. To avoid such disengagements, it is important to provide timely interventions. Such interventions are often expensive and can be provided to only a small fraction of the beneficiaries. We model this scenario as a restless multi-armed bandit (RMAB) problem, where each beneficiary is assumed to transition from one state to another depending on the intervention. Moreover, since the transition probabilities are unknown a priori, we propose a Whittle index based Q-Learning mechanism and show that it converges to the optimal solution. Our method improves over existing learning-based methods for RMABs on multiple benchmarks from literature and also on the maternal healthcare dataset. 2021-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6794 info:doi/10.24963/ijcai.2021/556 https://ink.library.smu.edu.sg/context/sis_research/article/7797/viewcontent/0556.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Planning and Scheduling Applications of Planning Artificial Intelligence and Robotics |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Planning and Scheduling Applications of Planning Artificial Intelligence and Robotics |
spellingShingle |
Planning and Scheduling Applications of Planning Artificial Intelligence and Robotics BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare |
description |
In many public health settings, it is important for patients to adhere to health programs, such as taking medications and periodic health checks. Unfortunately, beneficiaries may gradually disengage from such programs, which is detrimental to their health. A concrete example of gradual disengagement has been observed by an organization that carries out a free automated call-based program for spreading preventive care information among pregnant women. Many women stop picking up calls after being enrolled for a few months. To avoid such disengagements, it is important to provide timely interventions. Such interventions are often expensive and can be provided to only a small fraction of the beneficiaries. We model this scenario as a restless multi-armed bandit (RMAB) problem, where each beneficiary is assumed to transition from one state to another depending on the intervention. Moreover, since the transition probabilities are unknown a priori, we propose a Whittle index based Q-Learning mechanism and show that it converges to the optimal solution. Our method improves over existing learning-based methods for RMABs on multiple benchmarks from literature and also on the maternal healthcare dataset. |
format |
text |
author |
BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind |
author_facet |
BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind |
author_sort |
BISWAS, Arpita |
title |
Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare |
title_short |
Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare |
title_full |
Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare |
title_fullStr |
Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare |
title_full_unstemmed |
Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare |
title_sort |
learn to intervene: an adaptive learning policy for restless bandits in application to preventive healthcare |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2021 |
url |
https://ink.library.smu.edu.sg/sis_research/6794 https://ink.library.smu.edu.sg/context/sis_research/article/7797/viewcontent/0556.pdf |
_version_ |
1770576070168281088 |