Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Medical Informatics

Date Submitted: Dec 30, 2019
Date Accepted: Mar 19, 2020

The final, peer-reviewed published version of this preprint can be found here:

Document-Level Biomedical Relation Extraction Leveraging Pretrained Self-Attention Structure and Entity Replacement: Algorithm and Pretreatment Method Validation Study

Liu X, Fan J, Dong S

Document-Level Biomedical Relation Extraction Leveraging Pretrained Self-Attention Structure and Entity Replacement: Algorithm and Pretreatment Method Validation Study

JMIR Med Inform 2020;8(5):e17644

DOI: 10.2196/17644

PMID: 32469325

PMCID: 7314385

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Document-level Biomedical Relation Extraction Leveraging Pretrained Self-attention Structure and Entity Replacement: The Validation of Pretreatment Methods and Algorithm

  • Xiaofeng Liu; 
  • Jianye Fan; 
  • Shoubin Dong

ABSTRACT

Background:

The most current methods applied for the intra-sentence relation extraction in biomedical texts are inadequate for the document-level relation extraction, in which the relationship may cross sentence boundary. Hence, some approaches have been proposed to extract relations by splitting the document-level datasets through heuristic rules and learning methods. However, these approaches may introduce additional noise and do not really solve the problem of inter-sentence relation extraction. It is challenging to avoid noise and extract cross-sentence relation.

Objective:

This study is aimed to avoid errors by dividing the document-level dataset, to verify that self-attention structure can extract biomedical relation in a document with long distance dependencies and complex semantic, and to discuss which entity pretreatment method is better for biomedical relation extraction.

Methods:

The paper proposes a new data preprocessing and attempt to apply the pretrained self-attention structure for the document biomedical relation extraction with an entity replacement method to capture very long-distance dependencies and complex semantics.

Results:

Compared with the state-of-art approaches, our method greatly improved the value of precision. The result shows that our approach increases F1 value compared with the state-of-art methods. Through experiments of biomedical entity pretreatments, we find that model using an entity replacement method can improve performance.

Conclusions:

When considering all target entity pair as a whole in document-level dataset, pretrained self-attention structure is suitable to capture very long-distance dependencies and learn textual context and complicate semantics. Replacement method for biomedical entities is conductive to biomedical relation extraction, especially to document-level relation extraction.


 Citation

Please cite as:

Liu X, Fan J, Dong S

Document-Level Biomedical Relation Extraction Leveraging Pretrained Self-Attention Structure and Entity Replacement: Algorithm and Pretreatment Method Validation Study

JMIR Med Inform 2020;8(5):e17644

DOI: 10.2196/17644

PMID: 32469325

PMCID: 7314385

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.