Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Mental Health

Date Submitted: Oct 29, 2024
Open Peer Review Period: Oct 30, 2024 - Dec 25, 2024
Date Accepted: Nov 30, 2024
(closed for review but you can still tweet)

The final, peer-reviewed published version of this preprint can be found here:

Leveraging Large Language Models and Agent-Based Systems for Scientific Data Analysis: Validation Study

Peasley D, Kuplicki R, Sen S, Paulus M

Leveraging Large Language Models and Agent-Based Systems for Scientific Data Analysis: Validation Study

JMIR Ment Health 2025;12:e68135

DOI: 10.2196/68135

PMID: 39946556

PMCID: 11841814

Bridging Data and Insight: Leveraging LLMs and Agent-Based Systems for Scientific Data Analysis

  • Dale Peasley; 
  • Rayus Kuplicki; 
  • Sandip Sen; 
  • Martin Paulus

ABSTRACT

Background:

Large Language Models (LLMs) have shown promise in transforming how complex scientific data is analyzed and communicated, yet their application to scientific domains remains challenged by issues of factual accuracy and domain-specific precision. The LIBR-TU Research Agent (LITURAt) leverages a sophisticated agent-based architecture to mitigate these limitations, using external data retrieval and analysis tools to ensure reliable, context-aware outputs that make scientific information accessible to both experts and non-experts.

Objective:

The objective of this study was to develop and evaluate LITURAt to enable efficient analysis and contextualization of complex scientific datasets for diverse user expertise levels.

Methods:

An agent-based system based on LLMs was deviced to analyze and contextualize complex scientific datasets using a "plan-and-solve" framework. The system dynamically retrieves local data and relevant PubMed literature, performs statistical analyses, and generates comprehensive, context-aware summaries to answer user queries with high accuracy and consistency.

Results:

Our experiments demonstrated that LITURAt achieved an internal consistency rate of 94.8% and an external consistency rate of 91.9% across repeated and rephrased queries. Additionally, GPT-4 evaluations rated 80.3% of the system's answers as accurate and comprehensive, with 23.5% receiving the highest rating of 5 for completeness and precision.

Conclusions:

These findings highlight the potential of LITURAt to significantly enhance the accessibility and accuracy of scientific data analysis, achieving high consistency and strong performance in complex query resolution. Despite existing limitations, such as model stability for highly variable queries, LITURAt demonstrates promise as a robust tool for democratizing data-driven insights across diverse scientific domains.


 Citation

Please cite as:

Peasley D, Kuplicki R, Sen S, Paulus M

Leveraging Large Language Models and Agent-Based Systems for Scientific Data Analysis: Validation Study

JMIR Ment Health 2025;12:e68135

DOI: 10.2196/68135

PMID: 39946556

PMCID: 11841814

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.