Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Formative Research

Date Submitted: Jun 2, 2025
Open Peer Review Period: Jun 3, 2025 - Jun 3, 2025
Date Accepted: Dec 15, 2025
(closed for review but you can still tweet)

The final, peer-reviewed published version of this preprint can be found here:

Evaluating the Efficacy of AI-Based Interactive Assessments Using Large Language Models for Depression Screening: Development and Usability Study

Jin Z, Hu J, Zhou Q, Bi D, Zhao K, Yu H

Evaluating the Efficacy of AI-Based Interactive Assessments Using Large Language Models for Depression Screening: Development and Usability Study

JMIR Form Res 2026;10:e78401

DOI: 10.2196/78401

PMID: 41529832

PMCID: 12848484

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Efficacy of AI-Based Interactive Assessments Using Large Language Models for Depression Screening: Validation Study

  • Zheng Jin; 
  • Jiaxing Hu; 
  • Qian Zhou; 
  • Dandan Bi; 
  • Kaibin Zhao; 
  • Huan Yu

ABSTRACT

Background:

The evolution of language models, particularly Large Language Models (LLMs), has introduced transformative potential for psychological assessment, challenging traditional rating scale methods that have dominated clinical practice for over a century.

Objective:

This study aimed to develop and validate an Automated Assessment Paradigm (AAP) that integrates natural language processing (NLP) with conventional measurement tools to assess depressive symptoms, exploring its feasibility as a novel approach in psychological evaluation.

Methods:

A cohort of 115 participants, including 28 individuals diagnosed with depression based on the Diagnosis and Treatment of Mental Disorders Guidelines, completed the BDI-Fast Screen (BDI-FS) via a custom ChatGPT interface and the Chinese version of the PHQ-9. Statistical analyses included Spearman’s correlation (PHQ-9 vs. BDI-FS-GPT scores), Cohen’s kappa (diagnostic agreement), AUC evaluation, and logistic regression models comparing BDI-FS (cutoff = 3) and PHQ-9 (cutoff = 5).

Results:

Spearman's analysis revealed a significant correlation between PHQ-9 and BDI-FS-GPT total scores. Cohen’s kappa indicated substantial diagnostic agreement (κ = 0.75, P < .001). The BDI-FS demonstrated high diagnostic accuracy (AUC = 95.3%), detecting 89.3% of depressed patients with an 11.5% false-positive rate. Logistic regression models showed superior performance for BDI-FS compared to PHQ-9 in classification efficiency.

Conclusions:

The AAP framework combines the interactivity and personalization of NLP-powered tools with the psychometric rigor of traditional scales, suggesting a promising paradigm for future psychological assessment. Its ability to enhance engagement while maintaining reliability and validity supports further validation in large-scale studies as LLM technology advances.


 Citation

Please cite as:

Jin Z, Hu J, Zhou Q, Bi D, Zhao K, Yu H

Evaluating the Efficacy of AI-Based Interactive Assessments Using Large Language Models for Depression Screening: Development and Usability Study

JMIR Form Res 2026;10:e78401

DOI: 10.2196/78401

PMID: 41529832

PMCID: 12848484

The author of this paper has made a PDF available, but requires the user to login, or create an account.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.