Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Formative Research

Date Submitted: Nov 11, 2024
Date Accepted: Jan 5, 2025

The final, peer-reviewed published version of this preprint can be found here:

The Feasibility of Large Language Models in Verbal Comprehension Assessment: Mixed Methods Feasibility Study

Hadar-Shoval D, Lvovsky M, Asraf K, Shimoni Y, Elyoseph Z

The Feasibility of Large Language Models in Verbal Comprehension Assessment: Mixed Methods Feasibility Study

JMIR Form Res 2025;9:e68347

DOI: 10.2196/68347

PMID: 39993720

PMCID: 11894350

The Feasibility of Large Language Models in Verbal Comprehension Assessment: A Proof-of-Concept Study

  • Dorit Hadar-Shoval; 
  • Maya Lvovsky; 
  • Kfir Asraf; 
  • Yoav Shimoni; 
  • Zohar Elyoseph

ABSTRACT

Background:

Cognitive assessment is an important component of applied psychology, but limited access and high costs make these evaluations challenging.

Objective:

This pilot study examined the feasibility of using large language models (LLMs) to create personalized AI-based verbal comprehension tests (AI-BVCTs) for assessing verbal intelligence, in contrast with traditional assessment methods based on standardized norms.

Methods:

We used a within-subject design, comparing scores obtained from AI-BVCTs with those from the Wechsler Adult Intelligence Scale (WAIS-III) Verbal Comprehension Index (VCI).

Results:

The concordance correlation coefficient (CCC) demonstrated strong agreement between AI-BVCT and VCI scores (Claude: CCC = .752, 90% CI [.266, .933]; GPT-4: CCC = .733, 90% CI [.170, .935]). Pearson correlations further supported these findings, showing strong associations between VCI and AI-BVCT scores (Claude: r = .844, p < .001; GPT-4: r = .771, p = .025). No statistically significant differences were found between AI-BVCT and VCI scores (p > .05). These findings support the potential of LLMs to assess verbal intelligence.

Conclusions:

The study attests to the promise of AI-based cognitive tests in increasing the accessibility and affordability of assessment processes, enabling personalized testing. The research also raises ethical concerns regarding privacy and over-reliance on AI in clinical work. Further research with larger and more diverse samples is needed to establish the validity and reliability of this approach and develop more accurate scoring procedures.


 Citation

Please cite as:

Hadar-Shoval D, Lvovsky M, Asraf K, Shimoni Y, Elyoseph Z

The Feasibility of Large Language Models in Verbal Comprehension Assessment: Mixed Methods Feasibility Study

JMIR Form Res 2025;9:e68347

DOI: 10.2196/68347

PMID: 39993720

PMCID: 11894350

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.