Accepted for/Published in: Journal of Medical Internet Research
Date Submitted: Mar 18, 2022
Date Accepted: Jul 15, 2022
Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.
Predicting mortality in ICU Patients with heart failure using interpretable machine learning model
ABSTRACT
Background:
Heart failure (HF) is a common disease and a major public health problem. The HF mortality prediction is critical for developing individualized prevention and treatment plans. However, due to their lack of interpretability, most HF mortality prediction models have not yet reached clinical practice.
Objective:
We aimed to develop an interpretable model to predict the risk mortality for HF patients in intensive care unit (ICU) and use the Shapley additive explanation (SHAP) method to explain the extreme gradient boosting (XGBoost) model and explore prognostic factors for HF.
Methods:
In this retrospective cohort study, we achieved model development and performance comparison on the eICU Collaborative Research Database (eICU-CRD). We extracted data during the first 24 hours of each ICU admission and the data set was randomly divided, with 70% used for model training and 30% for model validation. The prediction performance of the XGBoost model was compared with three other machine learning models by the areas under the receiver operating characteristic curve (AUROC). Moreover, we used the Shapley additive explanation SHAP method to explain the XGBoost model.
Results:
A total of 2,798 eligible patients with HF were included in the final cohort for this study. The observed in-hospital mortality of patients with HF was 9.97%. Comparatively, the XGBoost model had the highest predictive performance among four models (AUC 0.824, 95% Confidence Interval (CI) 0.7766 to 0.8708), while support vector machine (SVM) had the poorest generalization ability (AUC 0.701, 95% CI 0.6433 to 0.7582). The decision curve showed that the net benefit of the XGBoost model surpassed those of other machine learning models at 10%~28% threshold probabilities. The SHAP method reveals the top 20 predictors of HF according to the importance ranking and the average of the blood urea nitrogen was recognized as the most important predictor variable.
Conclusions:
The interpretable predictive model helps physicians more accurately predict the mortality risk in ICU Patients with HF. This will help physicians to provide better treatment plan and optimal resource allocation for their patients. In addition, the interpretable framework can increase the transparency of the model and facilitates physicians to understand the reliability of the predictive model.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.