Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR AI

Date Submitted: Sep 5, 2024
Date Accepted: Jun 26, 2025

The final, peer-reviewed published version of this preprint can be found here:

Domain-Specific Pretraining of NorDeClin-Bidirectional Encoder Representations From Transformers for International Statistical Classification of Diseases, Tenth Revision, Code Prediction in Norwegian Clinical Texts: Model Development and Evaluation Study

Ngo PD, Tejedor Hernández MÃ, Chomutare T, Budrionis A, Svenning TO, Torsvik T, Lamproudis A, Dalianis H

Domain-Specific Pretraining of NorDeClin-Bidirectional Encoder Representations From Transformers for International Statistical Classification of Diseases, Tenth Revision, Code Prediction in Norwegian Clinical Texts: Model Development and Evaluation Study

JMIR AI 2025;4:e66153

DOI: 10.2196/66153

PMID: 40854226

PMCID: 12377785

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Continuous Pre-training and Performance Evaluation of NorDeClin-BERT for ICD-10 Coding Prediction

  • Phuong Dinh Ngo; 
  • Miguel Ãngel Tejedor Hernández; 
  • Taridzo Chomutare; 
  • Andrius Budrionis; 
  • Therese Olsen Svenning; 
  • Torbjørn Torsvik; 
  • Anastasios Lamproudis; 
  • Hercules Dalianis

ABSTRACT

Background:

Accurately assigning ICD-10 codes is critical for clinical documentation and epidemiological studies. Manual coding is time-consuming and prone to errors, underscoring the need for automated solutions in the Norwegian healthcare system.

Objective:

This study introduces NorDeClin-BERT, a Norwegian BERT-based model continuously pre-trained on a large corpus of Norwegian clinical text to predict ICD-10 diagnosis coding.

Methods:

NorDeClin-BERT was trained on the ClinCode Gastro Corpus, consisting of 8.8 million deidentified Norwegian clinical notes. The model’s performance was benchmarked against SweDeClin-BERT, SweClin-BERT, ScandiBERT, NorBERT3-base, and NorBERT3-large using accuracy, precision, recall, and F1-score metrics.

Results:

The results show that NorDeClin-BERT outperformed other similar-sized general BERT models and achieved equivalent performance with the larger NorBERT3-large model in classifying both prevalent and less common ICD-10 codes. The pretraining data, language specificity, and model architecture of NorDeClin-BERT contributed to its improved classification accuracy.

Conclusions:

This study highlights the potential of NorDeClin-BERT in improving the task of ICD-10 code classification for the gastroenterology domain in Norway, ultimately streamlining clinical documentation and reporting processes. The benchmarking evaluation establishes NorDeClin-BERT as a state-of-the-art model for processing Norwegian clinical text and predicting ICD-10 coding.


 Citation

Please cite as:

Ngo PD, Tejedor Hernández MÃ, Chomutare T, Budrionis A, Svenning TO, Torsvik T, Lamproudis A, Dalianis H

Domain-Specific Pretraining of NorDeClin-Bidirectional Encoder Representations From Transformers for International Statistical Classification of Diseases, Tenth Revision, Code Prediction in Norwegian Clinical Texts: Model Development and Evaluation Study

JMIR AI 2025;4:e66153

DOI: 10.2196/66153

PMID: 40854226

PMCID: 12377785

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.