Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Mental Health

Date Submitted: Nov 18, 2025
Open Peer Review Period: Nov 20, 2025 - Jan 15, 2026
Date Accepted: Feb 16, 2026
(closed for review but you can still tweet)

The final, peer-reviewed published version of this preprint can be found here:

Large Language Models and Their Applications in Mental Health: Scoping Review

Lokadjaja MC, Kho J, Goh WWB

Large Language Models and Their Applications in Mental Health: Scoping Review

JMIR Ment Health 2026;13:e88057

DOI: 10.2196/88057

PMID: 42139691

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Large Language Models and their Applications in Mental Health

  • Matheus Calvin Lokadjaja; 
  • Jordon Kho; 
  • Wilson Wen Bin Goh

ABSTRACT

Large language models (LLMs) are poised to transform mental healthcare, offering advanced capabilities in diagnosis, prognosis, and decision support. Since their inception, numerous mental health-focused LLMs have emerged in the scientific literature, reflecting the growing interest in leveraging these models across various clinical applications. With a broad range of models available, diverse tuning strategies, and multiple use cases, reviewing the current landscape is critical to understanding how LLMs are being applied. We screened 3,121 papers from PubMed, Scopus, and Web of Science focusing on model type and clinical use case. After removing duplicates and manual filtering, 42 studies were included in our final analysis. Most studies utilized OpenAI’s GPT series—GPT-4 (25 studies, 59.5%) and GPT-3.5 (16 studies, 38.1%) were the most common. Other frequently used models included BERT derived models (7 studies, 16.7%), LLaMA (8 studies, 18.6%), and RoBERTa derived models (6 studies, 14.0%). While all studies initially applied untuned LLMs, several adapted them through few-shot learning or fine-tuning to better align with specific research goals. Most models were used for diagnostic tasks (30 studies, 69.8%). The most common target conditions were depression (11 studies, 26.2%), followed by disorders such as ADHD, OCD, and suicidality. A subset of studies also examined general medical cases, which were included when mental health-related content was present. Despite rapid growth and diversity of LLM applications in mental health, the field remains nascent and exploratory. Future developments must emphasize responsible development, enhanced explainability, and deeper investigations into implementation and deployment practices centered on patient wellbeing.


 Citation

Please cite as:

Lokadjaja MC, Kho J, Goh WWB

Large Language Models and Their Applications in Mental Health: Scoping Review

JMIR Ment Health 2026;13:e88057

DOI: 10.2196/88057

PMID: 42139691

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.