Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Medical Informatics

Date Submitted: Mar 29, 2020
Date Accepted: Oct 28, 2020

The final, peer-reviewed published version of this preprint can be found here:

The Impact of Pretrained Language Models on Negation and Speculation Detection in Cross-Lingual Medical Text: Comparative Study

Rivera Zavala R, Martinez P

The Impact of Pretrained Language Models on Negation and Speculation Detection in Cross-Lingual Medical Text: Comparative Study

JMIR Med Inform 2020;8(12):e18953

DOI: 10.2196/18953

PMID: 33270027

PMCID: 7746498

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Analyzing the impact of pre-trained language models in negation and speculation detection in cross-lingual medical texts

  • Renzo Rivera Zavala; 
  • Paloma Martinez

ABSTRACT

Background:

Negation and speculation are critical elements in tasks related to Natural Language Processing, such as information extraction, as they change the truth-value of a proposition. In clinical narrative, these linguistic phenomena are used extensively with the objective of indicating hypothesis, impressions or negative findings. The previous state of the art approaches addressed negation and speculation detection tasks using rule-based methods but in the last few years, models based on machine learning and deep learning exploiting morphological, syntactic, semantic features represented as spare and dense vector, have emerged. However, although such methods of named entity recognition (NER) employ a broad set of features, they are limited to existing pre-trained models for a specific domain or language.

Objective:

A system for cross-lingual and cross-domain negation and speculation detection is introduced with special focus on biomedical scientific literature and clinical narrative. In this work, negation and speculation detection is considered as a sequence labeling task where cues and their scopes of both phenomena are recognized as a sequence of labels, recognized in an only phase.

Methods:

We propose two approaches: i) a Bidirectional Long Short-Term Memory (Bi-LSTM) and Conditional Random Field (CRF) using character, word and sense embeddings to deal with the extraction of semantic, syntactic and contextual patterns and ii) a Bidirectional Encoder Representations for Transformers (BERT) with fine-tuning for NER.

Results:

The approach was evaluated on English and Spanish in biomedical and reviews domains, particularly with the BioScope Corpus, IULA Corpus, and the SFU Spanish Review Corpus obtaining an F-measure of 86.6, 85.00, and 91.70, respectively.

Conclusions:

These results show that these architectures perform considerably better than the previous rule-based and machine learning-based systems. Moreover, our analysis results show that pre-training WordPieces and word embeddings on biomedical corpora help it to understand complexities inherent to biomedical texts.


 Citation

Please cite as:

Rivera Zavala R, Martinez P

The Impact of Pretrained Language Models on Negation and Speculation Detection in Cross-Lingual Medical Text: Comparative Study

JMIR Med Inform 2020;8(12):e18953

DOI: 10.2196/18953

PMID: 33270027

PMCID: 7746498

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.