Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: Journal of Medical Internet Research

Date Submitted: Mar 11, 2023
Date Accepted: Feb 27, 2024
Date Submitted to PubMed: Feb 29, 2024

The final, peer-reviewed published version of this preprint can be found here:

Evaluating Algorithmic Bias in 30-Day Hospital Readmission Models: Retrospective Analysis

Wang EH, Weiner JP, Saria S, Kharrazi H

Evaluating Algorithmic Bias in 30-Day Hospital Readmission Models: Retrospective Analysis

J Med Internet Res 2024;26:e47125

DOI: 10.2196/47125

PMID: 38422347

PMCID: 11066744

Algorithmic bias evaluation in 30-day hospital readmission models: A retrospective analysis of hospital discharges

  • Echo H. Wang; 
  • Jonathan P. Weiner; 
  • Suchi Saria; 
  • Hadi Kharrazi

ABSTRACT

Background:

The adoption of predictive algorithms in healthcare comes with the potential for algorithmic bias, which could exacerbate existing disparities. Fairness metrics were proposed to measure algorithm bias, but the application to real-world tasks is limited.

Objective:

This study aims to evaluate the algorithm bias associated with the application of common 30-day hospital readmission models and assesses the usefulness and interpretability of selected fairness metrics.

Methods:

This retrospective study used 10.6 million adult inpatient discharges from Maryland and Florida from 2016-2019. Models predicting 30-day hospital readmissions were evaluated: LACE Index, HOSPITAL score, and the CMS readmission measure, which was applied “as-is” (using existing coefficients) and “retrained” (recalibrated with 50% of the data). Predictive performances and bias measures were evaluated. Bias measures included the parity of false negative rate (FNR), false positive rate (FPR), zero-one-loss, and generalized entropy index.

Results:

The retrained CMS model and HOSPITAL score demonstrated the best overall predictive performance and the lowest racial and income bias in both states. In both states, white and higher income patient groups showed a higher FNR while Black and low-income patient groups resulted in a higher FPR and higher zero-one-loss. When stratified by hospital and population composition, these models demonstrated heterogenous algorithm bias.

Conclusions:

Caution must be taken when interpreting measures’ face value. A higher FNR or FPR could potentially reflect missed opportunities or wasted resources, but these measures could also reflect healthcare utilization patterns and gaps in care. The imperfections of health data, analytic frameworks, and the underlying health systems must be carefully considered. Fairness metrics can serve as a routine assessment to detect disparate model performances but are insufficient to inform mechanisms. Such assessment, however, is an important first step toward data-driven improvement to address existing health disparities.


 Citation

Please cite as:

Wang EH, Weiner JP, Saria S, Kharrazi H

Evaluating Algorithmic Bias in 30-Day Hospital Readmission Models: Retrospective Analysis

J Med Internet Res 2024;26:e47125

DOI: 10.2196/47125

PMID: 38422347

PMCID: 11066744

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.