Currently accepted at: JMIR Mental Health
Date Submitted: Nov 18, 2025
Open Peer Review Period: Nov 20, 2025 - Jan 15, 2026
Date Accepted: Feb 16, 2026
(closed for review but you can still tweet)
This paper has been accepted and is currently in production.
It will appear shortly on 10.2196/88057
The final accepted version (not copyedited yet) is in this tab.
Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.
Large Language Models and their Applications in Mental Health
ABSTRACT
Large language models (LLMs) are poised to transform mental healthcare, offering advanced capabilities in diagnosis, prognosis, and decision support. Since their inception, numerous mental health-focused LLMs have emerged in the scientific literature, reflecting the growing interest in leveraging these models across various clinical applications. With a broad range of models available, diverse tuning strategies, and multiple use cases, reviewing the current landscape is critical to understanding how LLMs are being applied. We screened 3,121 papers from PubMed, Scopus, and Web of Science focusing on model type and clinical use case. After removing duplicates and manual filtering, 42 studies were included in our final analysis. Most studies utilized OpenAI’s GPT series—GPT-4 (25 studies, 59.5%) and GPT-3.5 (16 studies, 38.1%) were the most common. Other frequently used models included BERT derived models (7 studies, 16.7%), LLaMA (8 studies, 18.6%), and RoBERTa derived models (6 studies, 14.0%). While all studies initially applied untuned LLMs, several adapted them through few-shot learning or fine-tuning to better align with specific research goals. Most models were used for diagnostic tasks (30 studies, 69.8%). The most common target conditions were depression (11 studies, 26.2%), followed by disorders such as ADHD, OCD, and suicidality. A subset of studies also examined general medical cases, which were included when mental health-related content was present. Despite rapid growth and diversity of LLM applications in mental health, the field remains nascent and exploratory. Future developments must emphasize responsible development, enhanced explainability, and deeper investigations into implementation and deployment practices centered on patient wellbeing.
Citation
Request queued. Please wait while the file is being generated. It may take some time.
Copyright
© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.