Interpretable machine learning models to predict hospital patient readmission
Background: Advanced machine learning models have received wide attention in assisting medical decision making due to the greater accuracy they can achieve. However, their limited interpretability imposes barriers for practitioners to adopt them. Recent advancements in interpretable machine learning...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/lkcsb_research/7666 https://ink.library.smu.edu.sg/context/lkcsb_research/article/8665/viewcontent/s12911_023_02193_5_pvoa_cc_by.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.lkcsb_research-8665 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.lkcsb_research-86652025-01-27T03:26:20Z Interpretable machine learning models to predict hospital patient readmission GAO, Xiaoquan ALAM, Sabriya SHI, Pengyi DEXTER, Franklin KONG, Nan Background: Advanced machine learning models have received wide attention in assisting medical decision making due to the greater accuracy they can achieve. However, their limited interpretability imposes barriers for practitioners to adopt them. Recent advancements in interpretable machine learning tools allow us to look inside the black box of advanced prediction methods to extract interpretable models while maintaining similar prediction accuracy, but few studies have investigated the specific hospital readmission prediction problem with this spirit. Methods: Our goal is to develop a machine-learning (ML) algorithm that can predict 30- and 90- day hospital readmissions as accurately as black box algorithms while providing medically interpretable insights into readmission risk factors. Leveraging a state-of-art interpretable ML model, we use a two-step Extracted Regression Tree approach to achieve this goal. In the first step, we train a black box prediction algorithm. In the second step, we extract a regression tree from the output of the black box algorithm that allows direct interpretation of medically relevant risk factors. We use data from a large teaching hospital in Asia to learn the ML model and verify our two-step approach. Results: The two-step method can obtain similar prediction performance as the best black box model, such as Neural Networks, measured by three metrics: accuracy, the Area Under the Curve (AUC) and the Area Under the Precision-Recall Curve (AUPRC), while maintaining interpretability. Further, to examine whether the prediction results match the known medical insights (i.e., the model is truly interpretable and produces reasonable results), we show that key readmission risk factors extracted by the two-step approach are consistent with those found in the medical literature. Conclusions: The proposed two-step approach yields meaningful prediction results that are both accurate and interpretable. This study suggests a viable means to improve the trust of machine learning based models in clinical practice for predicting readmissions through the two-step approach. 2023-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/lkcsb_research/7666 info:doi/10.1186/s12911-023-02193-5 https://ink.library.smu.edu.sg/context/lkcsb_research/article/8665/viewcontent/s12911_023_02193_5_pvoa_cc_by.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection Lee Kong Chian School Of Business eng Institutional Knowledge at Singapore Management University Hospital readmission Interpretable machine learning Risk prediction Administrative data Risk factors Health and Medical Administration Operations and Supply Chain Management |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Hospital readmission Interpretable machine learning Risk prediction Administrative data Risk factors Health and Medical Administration Operations and Supply Chain Management |
spellingShingle |
Hospital readmission Interpretable machine learning Risk prediction Administrative data Risk factors Health and Medical Administration Operations and Supply Chain Management GAO, Xiaoquan ALAM, Sabriya SHI, Pengyi DEXTER, Franklin KONG, Nan Interpretable machine learning models to predict hospital patient readmission |
description |
Background: Advanced machine learning models have received wide attention in assisting medical decision making due to the greater accuracy they can achieve. However, their limited interpretability imposes barriers for practitioners to adopt them. Recent advancements in interpretable machine learning tools allow us to look inside the black box of advanced prediction methods to extract interpretable models while maintaining similar prediction accuracy, but few studies have investigated the specific hospital readmission prediction problem with this spirit. Methods: Our goal is to develop a machine-learning (ML) algorithm that can predict 30- and 90- day hospital readmissions as accurately as black box algorithms while providing medically interpretable insights into readmission risk factors. Leveraging a state-of-art interpretable ML model, we use a two-step Extracted Regression Tree approach to achieve this goal. In the first step, we train a black box prediction algorithm. In the second step, we extract a regression tree from the output of the black box algorithm that allows direct interpretation of medically relevant risk factors. We use data from a large teaching hospital in Asia to learn the ML model and verify our two-step approach. Results: The two-step method can obtain similar prediction performance as the best black box model, such as Neural Networks, measured by three metrics: accuracy, the Area Under the Curve (AUC) and the Area Under the Precision-Recall Curve (AUPRC), while maintaining interpretability. Further, to examine whether the prediction results match the known medical insights (i.e., the model is truly interpretable and produces reasonable results), we show that key readmission risk factors extracted by the two-step approach are consistent with those found in the medical literature. Conclusions: The proposed two-step approach yields meaningful prediction results that are both accurate and interpretable. This study suggests a viable means to improve the trust of machine learning based models in clinical practice for predicting readmissions through the two-step approach. |
format |
text |
author |
GAO, Xiaoquan ALAM, Sabriya SHI, Pengyi DEXTER, Franklin KONG, Nan |
author_facet |
GAO, Xiaoquan ALAM, Sabriya SHI, Pengyi DEXTER, Franklin KONG, Nan |
author_sort |
GAO, Xiaoquan |
title |
Interpretable machine learning models to predict hospital patient readmission |
title_short |
Interpretable machine learning models to predict hospital patient readmission |
title_full |
Interpretable machine learning models to predict hospital patient readmission |
title_fullStr |
Interpretable machine learning models to predict hospital patient readmission |
title_full_unstemmed |
Interpretable machine learning models to predict hospital patient readmission |
title_sort |
interpretable machine learning models to predict hospital patient readmission |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/lkcsb_research/7666 https://ink.library.smu.edu.sg/context/lkcsb_research/article/8665/viewcontent/s12911_023_02193_5_pvoa_cc_by.pdf |
_version_ |
1823108751291318272 |