Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Medical Informatics

Date Submitted: May 3, 2024
Date Accepted: Aug 6, 2024

The final, peer-reviewed published version of this preprint can be found here:

Health Care Language Models and Their Fine-Tuning for Information Extraction: Scoping Review

Nunes M, Bone J, Ferreira JC, Elvas LB

Health Care Language Models and Their Fine-Tuning for Information Extraction: Scoping Review

JMIR Med Inform 2024;12:e60164

DOI: 10.2196/60164

PMID: 39432345

PMCID: 11535799

Healthcare Language Models and their fine-tuning for Information Extraction: Systematic Review

  • Miguel Nunes; 
  • Joao Bone; 
  • Joao C Ferreira; 
  • Luis B. Elvas

ABSTRACT

Background:

In response to the intricate language and specialized terminology inherent in healthcare text data, domain adaptation techniques have emerged as crucial to Transformer-based models for improved performance in downstream tasks, such as Information Extraction.

Objective:

This study presents a systematic literature review investigating domain adaptation methods for Transformers in healthcare differentiating between English and non-English languages, with a focus on Portuguese, addressing respective language-specific challenges, and information extraction models to assess the efficacy of Transformer-based approaches in healthcare contexts.

Methods:

This systematic review was conducted using the Preferred Reporting Items for Systematic Reviews and Meta-analyzes (PRISMA) methodology on Scopus and Web of Science Core Collection databases. Only studies that mentioned the creation of healthcare language models or healthcare information extraction models were included, while large language models were excluded.

Results:

Our search query retrieved 137 articles, 60 of which met the inclusion criteria. Furthermore, 5 studies were included via other methods. For studies mentioning the creation of healthcare language models, 7 were trained using English data, 4 using Chinese data, and only 3 using Portuguese, with 2 of them being Brazilian Portuguese. Regarding Information Extraction models, Transformers were the most commonly utilized method, with 16 articles, while other Deep Learning or Machine Learning methods were also popular. Nevertheless, the current state-of-the-art for Information Extraction models involves utilizing Transformers. The most extracted entities were diagnosis, posology, symptoms, and phenotype related to specific diseases.

Conclusions:

The findings indicate that domain adaptation is beneficial for the healthcare context, achieving better results in downstream tasks. Our analysis allowed us to understand that the utilization of Transformers is more developed for the English language, while non-English languages face challenges regarding the lack of available data, resources, or relevant studies, with the Chinese language having the most results. The Portuguese language lacks relevant studies in its European version and should draw examples from other non-English languages to develop these models and drive progress in AI. For the Information Extraction task, several methods were employed, but the current state-of-the-art utilizes Transformers.


 Citation

Please cite as:

Nunes M, Bone J, Ferreira JC, Elvas LB

Health Care Language Models and Their Fine-Tuning for Information Extraction: Scoping Review

JMIR Med Inform 2024;12:e60164

DOI: 10.2196/60164

PMID: 39432345

PMCID: 11535799

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.