Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.
Comparing Deep Learning Approaches for Predicting Clinical Deterioration Using Chest Radiographs
ABSTRACT
Background:
Early detection of clinical deterioration and timely intervention for hospitalized patients can improve patient outcomes. Existing early warning systems rely on variables from structured data, such as vital signs and laboratory values, and do not incorporate other potentially predictive data modalities. Because respiratory failure is a common cause of deterioration, chest radiographs are often acquired in deteriorating patients, which may be informative for predicting their risk of intensive care unit (ICU) transfer.
Objective:
To compare and validate different computer vision models and data augmentation approaches with chest radiographs for predicting clinical deterioration.
Methods:
This retrospective observational study included adult patients hospitalized at the University of Wisconsin Health System between 2009 and 2020 with an elevated eCART score, a validated clinical deterioration early warning score, on the medical-surgical wards. Patients with a chest radiograph within 48 hours prior to the elevated score were included in this study. Three computer vision model architectures (VGG16, Densenet121, Vision Transformer) and four data augmentation methods (Histogram Normalization, Random Flip, Random Gaussian Noise, and Random Rotate) were compared using the area under the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AUPRC) for predicting clinical deterioration (i.e., intensive care unit transfer or ward death in the following 24 hours).
Results:
The study included 21,817 patient admissions, of which 1,655 (7.6%) experienced the outcome. Densenet121 model pre-trained on chest radiograph datasets with histogram normalization and random Gaussian noise augmentation had the highest discrimination (AUROC 0.734 and AUPRC 0.414), while vision transformer having 24 transformer blocks with random rotate augmentation had the lowest discrimination (AUROC=0.598).
Conclusions:
The Densenet121 architecture pretrained with chest radiographs performed better than other architectures in most experiments, and the addition of histogram normalization with random Gaussian noise data augmentation may enhance performance for Densenet121 and pre-trained VGG16 architectures.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.