Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Formative Research

Date Submitted: Feb 23, 2024
Date Accepted: May 6, 2025

The final, peer-reviewed published version of this preprint can be found here:

Comparative Efficacy of MultiModal AI Methods in Screening for Major Depressive Disorder: Machine Learning Model Development Predictive Pilot Study

Chen D, Wang P, Zhang X, Qiao R, Li N, Zhang X, Zhang H, Wang G

Comparative Efficacy of MultiModal AI Methods in Screening for Major Depressive Disorder: Machine Learning Model Development Predictive Pilot Study

JMIR Form Res 2025;9:e56057

DOI: 10.2196/56057

PMID: 40446148

PMCID: 12143584

Comparative Efficacy of Multi-Modal AI Methods in Screening for Major Depressive Disorder: Machine Learning Predictive Pilot Study

  • Donghao Chen; 
  • Pengfei Wang; 
  • Xiaolong Zhang; 
  • Runqi Qiao; 
  • Nanxi Li; 
  • Xiaodong Zhang; 
  • Honggang Zhang; 
  • Gang Wang

ABSTRACT

Background:

With an increasing number of Major Depressive Disorder (MDD) patients, an efficient and accurate tool for screening and auxiliary diagnosis is in need. This process is often done by mental status examination and assessment scales. In recent years, Artificial Intelligence (AI) has been used on some auxiliary diagnosis such as pulmonary nodules diagnose. We want to use AI to help us to do the screening and auxiliary diagnosis, explore effective stimulus form and comprehensive makers.

Objective:

To find an AI method to do the screening and auxiliary diagnosis of MDD on videos based on specially designed multi-stimulus diagram, which is not scale way, and find effective stimulus and comprehensive markers.

Methods:

We designed a novel way to make the prediction, which included questionnaire, question and answering (Q&A), mental imagery description and video watching. A total of 89 persons participated in our project. We extracted multi-modal features which included vision, audio and text and trained a deep-learning based classification model to give the screening and extent estimation. We trained models for stimulus combinations and got the corresponding performance to evaluate the effectiveness. We used visualization method to see what the AI models learn and compare with correlation analysis results to find valuable markers.

Results:

Top screening accuracy and extent accuracy was 94.12% and 76.47% respectively. Q&A performed best in the listed stimuli. Based on the comparison between the attention learned and traditional correlation results, the agreement exists and there were some typical symptoms of each group.

Conclusions:

In the field of computational psychiatry, deep learning based multimodal-analyzing can help to do the screening and auxiliary diagnosis of MDD, depression related Q&A is the most effective stimulus and AI learned attention which is consistent to human’s prior experience.


 Citation

Please cite as:

Chen D, Wang P, Zhang X, Qiao R, Li N, Zhang X, Zhang H, Wang G

Comparative Efficacy of MultiModal AI Methods in Screening for Major Depressive Disorder: Machine Learning Model Development Predictive Pilot Study

JMIR Form Res 2025;9:e56057

DOI: 10.2196/56057

PMID: 40446148

PMCID: 12143584

Per the author's request the PDF is not available.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.