Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: Journal of Medical Internet Research

Date Submitted: Aug 5, 2024
Date Accepted: Sep 27, 2024

The final, peer-reviewed published version of this preprint can be found here:

Authors’ Reply: Reassessing AI in Medicine: Exploring the Capabilities of AI in Academic Abstract Synthesis

Hsu TW, Liang CS

Authors’ Reply: Reassessing AI in Medicine: Exploring the Capabilities of AI in Academic Abstract Synthesis

J Med Internet Res 2024;26:e65123

DOI: 10.2196/65123

PMID: 39715539

PMCID: 11704654

Reply to: Reassessing AI in Medicine: Exploring the Capabilities of AI in Academic Abstract Synthesis

  • Tien-Wei Hsu; 
  • Chih-Sung Liang

ABSTRACT

We appreciate Mr. Wang and his colleague's keen interest in our publication [1]. The responses to his comments are as follows. First, Mr. Wang and his colleague argued potential biases from the eight experts' personal preferences, academic or cultural backgrounds, or their familiarity with the subject matter. However, the 30 papers were selected from three different journals and explored various topics (Table S3 in the supplementary data). Therefore, these biases are unlikely to occur due to the significant variation among the articles. Second, we did not use ChatGPT 3.5. We used ChatPDF, which is based on ChatGPT 3.5 but can efficiently analyze all PDF contents. Third, Mr. Wang and his colleague argued that if AI assistance in refining an article does not alter the accuracy of the research and aids in clearer communication of experimental content, such practice should be deemed acceptable. However, in our study, we found that 3 of the 30 ChatGPT-generated abstracts showed wrong conclusions. Finally, Mr. Wang and his colleague are right in cautioning against complete reliance on GPT for labor-intensive tasks, which may lead to a lack of critical thinking and the production of low-quality or erroneous articles. Indeed, when using GPTzero [2] to check their letter, the probability of AI-written text is 81%, not 100%. However, their perspective differs from the purpose of our study. We aimed to assess the applicability of an AI model in generating abstracts for basic preclinical research. Our conclusion is that the quality of the ChatGPT-generated abstracts of basic preclinical research was suboptimal, and the accuracy was not 100%.


 Citation

Please cite as:

Hsu TW, Liang CS

Authors’ Reply: Reassessing AI in Medicine: Exploring the Capabilities of AI in Academic Abstract Synthesis

J Med Internet Res 2024;26:e65123

DOI: 10.2196/65123

PMID: 39715539

PMCID: 11704654

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.