Accepted for/Published in: JMIR Formative Research
Date Submitted: Mar 22, 2024
Open Peer Review Period: Sep 25, 2024 - Nov 20, 2024
Date Accepted: Oct 10, 2024
(closed for review but you can still tweet)
Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.
Does facilitating trust calibration for artificial-intelligence-driven differential diagnoses list improve physicians' diagnostic accuracy?: A quasi-experimental study
ABSTRACT
Background:
Despite the usefulness of artificial intelligence (AI)-based diagnostic decision-support systems, the over-reliance of physicians on AI-generated diagnoses may lead to diagnostic errors.
Objective:
We investigated the safe use of AI-based diagnostic-support systems with trust calibration, adjusting trust levels to AI’s actual reliability.
Methods:
A quasi-experimental study was conducted at Dokkyo Medical University, Japan, with physicians allocated (1:1) to the intervention and control groups. The participants reviewed medical histories of 20 clinical cases generated by an AI-driven automated medical history-taking system with an AI-generated list of 10 differential diagnoses and provided one to three possible diagnoses. Physicians were asked to consider whether the final diagnosis was included in the AI-generated list of 10 differential diagnoses in the intervention group, which served as trust calibration. We analyzed the diagnostic accuracy of physicians and the correctness of trust calibration in the intervention group.
Results:
Among the 20 physicians assigned to the intervention (n=10) and control (n=10) groups, the diagnostic accuracy was 41.5% and 46.0%, respectively, without significant difference (odds ratio 1.20, 95% confidence interval [CI] 0.81–1.78, P=.42). The overall accuracy of the trust calibration was only 61.5%, and despite correct calibration, the diagnostic accuracy was 54.5%.
Conclusions:
Trust calibration did not significantly improve physicians' diagnostic accuracy when considering differential diagnoses generated by reading medical histories and the possible differential diagnosis lists of an AI-driven automated medical history-taking system. This study underscores the limitations of the extant trust-calibration system and highlights the need to apply supportive measures of trust calibration rather than solely utilizing trust calibration.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.