Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Currently accepted at: JMIR Mental Health

Date Submitted: Jun 1, 2025
Date Accepted: Feb 27, 2026

This paper has been accepted and is currently in production.

It will appear shortly on 10.2196/78351

The final accepted version (not copyedited yet) is in this tab.

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Errors in Generative AI-Based Patient-Centered Mental Health Documentation: Psychiatrists’ Qualitative Pre-Post Comparison

  • Pelin Ozkara Menekseoglu; 
  • Mareike Weibezahl; 
  • Mats Ellingsen; 
  • Jarl Sterkenburg; 
  • Anna Kharko; 
  • Stefan Hochwarter; 
  • Julian Schwarz

ABSTRACT

Background:

Patients' digital access to their personal health data is becoming increasingly common worldwide. However, medical documentation often contains technical language and sensitive information, which can lead to potential misunderstandings and distress among patients. These issues may be particularly impactful in mental health contexts. Large language models (LLMs) offer a promising approach by transforming clinician-generated health notes into language that is more patient-centered, non medicalized, and empathetic. However, risks related to accuracy and clinical safety have not been adequately investigated in psychiatry.

Objective:

To qualitatively analyze the errors introduced by LLMs when transforming notes written by psychiatrists into patient-facing formats. The study will also highlight the implications for clinical communication and patient safety.

Methods:

Clinical notes (n = 63) written by 19 psychiatrists in an outpatient treatment setting were collected, anonymized, and translated from German to English by humans. OpenAI's GPT-3.5 Turbo was used to develop a pre-prompt that transformed these notes into a patient-centered, lay-readable form through an iterative process. Three psychiatrists qualitatively analyzed the LLM-revised documentation using Kuckartz content analysis. They compared the pre- and post-conversion notes to systematically identify and categorize LLM-induced errors.

Results:

Five categories of clinically relevant errors were identified: (1) clinical misinterpretations, particularly in critical assessments such as suicidality, where nuanced terminology was oversimplified or inaccurately represented; (2) attribution errors, where behaviors or roles within family dynamics or interactions were incorrectly attributed to different individuals; (3) content distortion errors were characterized by speculative additions, emotional exaggerations, and inappropriate contextual assumptions; (4) abbreviation and terminology errors resulted from inaccurate expansions of medical abbreviations and terms; and (5) structural and syntax errors resulted in ambiguity, particularly when the original notes were brief or bulleted. Despite significant improvements in the readability and overall linguistic fluency of the converted notes, these errors occurred.

Conclusions:

LLMs have the potential to transform psychiatric notes into patient-friendly formats. However, critical errors remain prevalent and can impair clinical judgment, understanding of patient circumstances, clarity of medication regimens, and interpretation of clinical observations. To safely integrate AI-generated documentation into psychiatric care, clinician oversight and targeted model refinement are essential. Future research should explore strategies to mitigate these errors, assess their comprehensive clinical impact, and incorporate patient and provider perspectives to ensure robust implementation. Clinical Trial: German Clinical Trial Register, Registiration Number: DRKS00030188, URL: https://drks.de/search/en/trial/DRKS00030188


 Citation

Please cite as:

Ozkara Menekseoglu P, Weibezahl M, Ellingsen M, Sterkenburg J, Kharko A, Hochwarter S, Schwarz J

Errors in Generative AI-Based Patient-Centered Mental Health Documentation: Psychiatrists’ Qualitative Pre-Post Comparison

JMIR Preprints. 01/06/2025:78351

DOI: 10.2196/preprints.78351

URL: https://preprints.jmir.org/preprint/78351

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.