Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare

In many public health settings, it is important for patients to adhere to health programs, such as taking medications and periodic health checks. Unfortunately, beneficiaries may gradually disengage from such programs, which is detrimental to their health. A concrete example of gradual disengagement...

全面介紹

Saved in:
書目詳細資料
Main Authors: BISWAS, Arpita, AGGARWAL, Gaurav, VARAKANTHAM, Pradeep, TAMBE, Milind
格式: text
語言:English
出版: Institutional Knowledge at Singapore Management University 2021
主題:
在線閱讀:https://ink.library.smu.edu.sg/sis_research/6794
https://ink.library.smu.edu.sg/context/sis_research/article/7797/viewcontent/0556.pdf
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
id sg-smu-ink.sis_research-7797
record_format dspace
spelling sg-smu-ink.sis_research-77972022-01-27T09:57:40Z Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare BISWAS, Arpita AGGARWAL, Gaurav VARAKANTHAM, Pradeep TAMBE, Milind In many public health settings, it is important for patients to adhere to health programs, such as taking medications and periodic health checks. Unfortunately, beneficiaries may gradually disengage from such programs, which is detrimental to their health. A concrete example of gradual disengagement has been observed by an organization that carries out a free automated call-based program for spreading preventive care information among pregnant women. Many women stop picking up calls after being enrolled for a few months. To avoid such disengagements, it is important to provide timely interventions. Such interventions are often expensive and can be provided to only a small fraction of the beneficiaries. We model this scenario as a restless multi-armed bandit (RMAB) problem, where each beneficiary is assumed to transition from one state to another depending on the intervention. Moreover, since the transition probabilities are unknown a priori, we propose a Whittle index based Q-Learning mechanism and show that it converges to the optimal solution. Our method improves over existing learning-based methods for RMABs on multiple benchmarks from literature and also on the maternal healthcare dataset. 2021-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6794 info:doi/10.24963/ijcai.2021/556 https://ink.library.smu.edu.sg/context/sis_research/article/7797/viewcontent/0556.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Planning and Scheduling Applications of Planning Artificial Intelligence and Robotics
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Planning and Scheduling
Applications of Planning
Artificial Intelligence and Robotics
spellingShingle Planning and Scheduling
Applications of Planning
Artificial Intelligence and Robotics
BISWAS, Arpita
AGGARWAL, Gaurav
VARAKANTHAM, Pradeep
TAMBE, Milind
Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
description In many public health settings, it is important for patients to adhere to health programs, such as taking medications and periodic health checks. Unfortunately, beneficiaries may gradually disengage from such programs, which is detrimental to their health. A concrete example of gradual disengagement has been observed by an organization that carries out a free automated call-based program for spreading preventive care information among pregnant women. Many women stop picking up calls after being enrolled for a few months. To avoid such disengagements, it is important to provide timely interventions. Such interventions are often expensive and can be provided to only a small fraction of the beneficiaries. We model this scenario as a restless multi-armed bandit (RMAB) problem, where each beneficiary is assumed to transition from one state to another depending on the intervention. Moreover, since the transition probabilities are unknown a priori, we propose a Whittle index based Q-Learning mechanism and show that it converges to the optimal solution. Our method improves over existing learning-based methods for RMABs on multiple benchmarks from literature and also on the maternal healthcare dataset.
format text
author BISWAS, Arpita
AGGARWAL, Gaurav
VARAKANTHAM, Pradeep
TAMBE, Milind
author_facet BISWAS, Arpita
AGGARWAL, Gaurav
VARAKANTHAM, Pradeep
TAMBE, Milind
author_sort BISWAS, Arpita
title Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
title_short Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
title_full Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
title_fullStr Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
title_full_unstemmed Learn to intervene: An adaptive learning policy for restless bandits in application to preventive healthcare
title_sort learn to intervene: an adaptive learning policy for restless bandits in application to preventive healthcare
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/6794
https://ink.library.smu.edu.sg/context/sis_research/article/7797/viewcontent/0556.pdf
_version_ 1770576070168281088