Interpretable machine learning models to predict hospital patient readmission

Background: Advanced machine learning models have received wide attention in assisting medical decision making due to the greater accuracy they can achieve. However, their limited interpretability imposes barriers for practitioners to adopt them. Recent advancements in interpretable machine learning...

Full description

Saved in:
Bibliographic Details
Main Authors: GAO, Xiaoquan, ALAM, Sabriya, SHI, Pengyi, DEXTER, Franklin, KONG, Nan
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/lkcsb_research/7666
https://ink.library.smu.edu.sg/context/lkcsb_research/article/8665/viewcontent/s12911_023_02193_5_pvoa_cc_by.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.lkcsb_research-8665
record_format dspace
spelling sg-smu-ink.lkcsb_research-86652025-01-27T03:26:20Z Interpretable machine learning models to predict hospital patient readmission GAO, Xiaoquan ALAM, Sabriya SHI, Pengyi DEXTER, Franklin KONG, Nan Background: Advanced machine learning models have received wide attention in assisting medical decision making due to the greater accuracy they can achieve. However, their limited interpretability imposes barriers for practitioners to adopt them. Recent advancements in interpretable machine learning tools allow us to look inside the black box of advanced prediction methods to extract interpretable models while maintaining similar prediction accuracy, but few studies have investigated the specific hospital readmission prediction problem with this spirit. Methods: Our goal is to develop a machine-learning (ML) algorithm that can predict 30- and 90- day hospital readmissions as accurately as black box algorithms while providing medically interpretable insights into readmission risk factors. Leveraging a state-of-art interpretable ML model, we use a two-step Extracted Regression Tree approach to achieve this goal. In the first step, we train a black box prediction algorithm. In the second step, we extract a regression tree from the output of the black box algorithm that allows direct interpretation of medically relevant risk factors. We use data from a large teaching hospital in Asia to learn the ML model and verify our two-step approach. Results: The two-step method can obtain similar prediction performance as the best black box model, such as Neural Networks, measured by three metrics: accuracy, the Area Under the Curve (AUC) and the Area Under the Precision-Recall Curve (AUPRC), while maintaining interpretability. Further, to examine whether the prediction results match the known medical insights (i.e., the model is truly interpretable and produces reasonable results), we show that key readmission risk factors extracted by the two-step approach are consistent with those found in the medical literature. Conclusions: The proposed two-step approach yields meaningful prediction results that are both accurate and interpretable. This study suggests a viable means to improve the trust of machine learning based models in clinical practice for predicting readmissions through the two-step approach. 2023-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/lkcsb_research/7666 info:doi/10.1186/s12911-023-02193-5 https://ink.library.smu.edu.sg/context/lkcsb_research/article/8665/viewcontent/s12911_023_02193_5_pvoa_cc_by.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection Lee Kong Chian School Of Business eng Institutional Knowledge at Singapore Management University Hospital readmission Interpretable machine learning Risk prediction Administrative data Risk factors Health and Medical Administration Operations and Supply Chain Management
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Hospital readmission
Interpretable machine learning
Risk prediction
Administrative data
Risk factors
Health and Medical Administration
Operations and Supply Chain Management
spellingShingle Hospital readmission
Interpretable machine learning
Risk prediction
Administrative data
Risk factors
Health and Medical Administration
Operations and Supply Chain Management
GAO, Xiaoquan
ALAM, Sabriya
SHI, Pengyi
DEXTER, Franklin
KONG, Nan
Interpretable machine learning models to predict hospital patient readmission
description Background: Advanced machine learning models have received wide attention in assisting medical decision making due to the greater accuracy they can achieve. However, their limited interpretability imposes barriers for practitioners to adopt them. Recent advancements in interpretable machine learning tools allow us to look inside the black box of advanced prediction methods to extract interpretable models while maintaining similar prediction accuracy, but few studies have investigated the specific hospital readmission prediction problem with this spirit. Methods: Our goal is to develop a machine-learning (ML) algorithm that can predict 30- and 90- day hospital readmissions as accurately as black box algorithms while providing medically interpretable insights into readmission risk factors. Leveraging a state-of-art interpretable ML model, we use a two-step Extracted Regression Tree approach to achieve this goal. In the first step, we train a black box prediction algorithm. In the second step, we extract a regression tree from the output of the black box algorithm that allows direct interpretation of medically relevant risk factors. We use data from a large teaching hospital in Asia to learn the ML model and verify our two-step approach. Results: The two-step method can obtain similar prediction performance as the best black box model, such as Neural Networks, measured by three metrics: accuracy, the Area Under the Curve (AUC) and the Area Under the Precision-Recall Curve (AUPRC), while maintaining interpretability. Further, to examine whether the prediction results match the known medical insights (i.e., the model is truly interpretable and produces reasonable results), we show that key readmission risk factors extracted by the two-step approach are consistent with those found in the medical literature. Conclusions: The proposed two-step approach yields meaningful prediction results that are both accurate and interpretable. This study suggests a viable means to improve the trust of machine learning based models in clinical practice for predicting readmissions through the two-step approach.
format text
author GAO, Xiaoquan
ALAM, Sabriya
SHI, Pengyi
DEXTER, Franklin
KONG, Nan
author_facet GAO, Xiaoquan
ALAM, Sabriya
SHI, Pengyi
DEXTER, Franklin
KONG, Nan
author_sort GAO, Xiaoquan
title Interpretable machine learning models to predict hospital patient readmission
title_short Interpretable machine learning models to predict hospital patient readmission
title_full Interpretable machine learning models to predict hospital patient readmission
title_fullStr Interpretable machine learning models to predict hospital patient readmission
title_full_unstemmed Interpretable machine learning models to predict hospital patient readmission
title_sort interpretable machine learning models to predict hospital patient readmission
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/lkcsb_research/7666
https://ink.library.smu.edu.sg/context/lkcsb_research/article/8665/viewcontent/s12911_023_02193_5_pvoa_cc_by.pdf
_version_ 1823108751291318272