Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR AI

Date Submitted: Aug 27, 2023
Date Accepted: Feb 3, 2024

The final, peer-reviewed published version of this preprint can be found here:

The Impact of Expectation Management and Model Transparency on Radiologists’ Trust and Utilization of AI Recommendations for Lung Nodule Assessment on Computed Tomography: Simulated Use Study

Ewals LJ, Heesterbeek LJ, Yu B, van der Wulp K, Mavroeidis D, Funk M, Snijders CC, Jacobs I, Nederend J, Pluyter JR, e/MTIC Oncology group

The Impact of Expectation Management and Model Transparency on Radiologists’ Trust and Utilization of AI Recommendations for Lung Nodule Assessment on Computed Tomography: Simulated Use Study

JMIR AI 2024;3:e52211

DOI: 10.2196/52211

PMID: 38875574

PMCID: 11041414

Appropriate trust in AI: a pilot study on the impact of expectation management and model transparency on radiologists’ trust and utilization of AI recommendations for lung nodule assessment on CT

  • Lotte J.S. Ewals; 
  • Lynn J.J. Heesterbeek; 
  • Bin Yu; 
  • Kasper van der Wulp; 
  • Dimitrios Mavroeidis; 
  • Mathias Funk; 
  • Cris C.P. Snijders; 
  • Igor Jacobs; 
  • Joost Nederend; 
  • Jon R. Pluyter; 
  • e/MTIC Oncology group

ABSTRACT

Background:

Many promising Artificial Intelligence (AI) and Computer-Aided Detection and Diagnosis (CAD) systems have been developed in lab settings, but few have been successfully integrated into clinical practice. This is partially due to a lack of user-centered design of AI-CAD systems.

Objective:

To assess the impact of different onboarding tutorials and different levels of AI model explainability on radiologists’ trust in AI and the utilization of AI recommendations in lung nodule assessment on CT scans.

Methods:

Twenty radiologists from 7 Dutch medical centers performed lung nodule assessment on CT scans under different conditions in a simulated use study as part of a 2x2 repeated measures experimental design. Two types of AI onboarding tutorials (reflective vs. informative) and two levels of AI output (black box vs. explainable) were designed. Participating radiologists first received an onboarding tutorial, which was either informative or reflective. Subsequently, each radiologist assessed 7 CT scans, first without AI recommendations. AI recommendations were shown to the radiologist, and they could adjust their initial assessment if they found it appropriate. Half of the participants received the recommendations via black box AI output and the other half received explainable AI output. Mental model and psychological trust were measured before onboarding, after onboarding and after assessing the 7 CT scans. We recorded whether radiologists changed their assessment on found nodules, their malignancy prediction, and follow-up advice for each CT assessment. In addition, we analyzed whether radiologists’ trust in their assessments had changed based on the AI recommendations.

Results:

Both variations of onboarding tutorials resulted in a significantly better understanding of the capabilities and limitations of the AI-CAD system (informative p=0.011 and reflective p=0.007). After using AI-CAD, the psychological trust significantly decreased for the group with explainable AI output (p=0.021). Based on the AI recommendations, radiologists changed the number of reported nodules in 27/140 assessments, the malignancy prediction in 32/140 assessments and the follow-up advice in 12/140 assessments. The changes were mostly an increased number of reported nodules, a higher estimated probability of malignancy, and earlier follow-up. The radiologists’ confidence in their found nodules changed in 82/140 assessments, in their estimated probability of malignancy in 50/140 assessments, and in their follow-up advice in 28/140 assessments. These changes were predominantly increases in confidence. The number of changed assessments and changed radiologists’ confidence did not significantly differ between the groups that received different onboarding tutorials and AI outputs.

Conclusions:

Onboarding tutorials help radiologists in getting a better understanding of the AI-CAD and facilitate formation of a correct mental model of the AI-CAD. If AI explanations do not consistently substantiate the probability of malignancy across patient cases, this can impair radiologists' trust in the AI-CAD system. Radiologists’ confidence in their assessments was improved by using the AI recommendations. Clinical Trial: NA


 Citation

Please cite as:

Ewals LJ, Heesterbeek LJ, Yu B, van der Wulp K, Mavroeidis D, Funk M, Snijders CC, Jacobs I, Nederend J, Pluyter JR, e/MTIC Oncology group

The Impact of Expectation Management and Model Transparency on Radiologists’ Trust and Utilization of AI Recommendations for Lung Nodule Assessment on Computed Tomography: Simulated Use Study

JMIR AI 2024;3:e52211

DOI: 10.2196/52211

PMID: 38875574

PMCID: 11041414

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.