Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR mHealth and uHealth

Date Submitted: Feb 1, 2022
Date Accepted: Mar 21, 2022

The final, peer-reviewed published version of this preprint can be found here:

Fully Automated Wound Tissue Segmentation Using Deep Learning on Mobile Devices: Cohort Study

Ramachandram D, Ramirez-GarciaLuna J, Fraser RD, Martinez-Jimenez M, Arriaga-Caballero J, Allport J

Fully Automated Wound Tissue Segmentation Using Deep Learning on Mobile Devices: Cohort Study

JMIR Mhealth Uhealth 2022;10(4):e36977

DOI: 10.2196/36977

PMID: 35451982

PMCID: 9077502

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Improving Objective Wound Assessment: Fully-automated wound tissue segmentation using Deep Learning on mobile devices

  • Dhanesh Ramachandram; 
  • Jose Ramirez-GarciaLuna; 
  • Robert D.J. Fraser; 
  • Mario Martinez-Jimenez; 
  • Jesus Arriaga-Caballero; 
  • Justin Allport

ABSTRACT

Background:

Composition of tissue types present within a wound is an useful indicator of its healing progression and could be helpful in guiding its treatment. Additionally, this measure is clinically used in wound healing tools (e.g. BWAT) to assess risk and recommend treatment. However, the identification of wound tissue and the estimation of their relative composition is highly subjective and variable. This results in incorrect assessments being reported, leading to downstream impacts including inappropriate dressing selection, failure to identify wounds at risk of not healing, or failure to make appropriate referrals to specialists.

Objective:

To measure inter- and intra-rater variability in manual tissue segmentation and quantification among a cohort of wound care clinicians. To determine if an objective assessment of tissue types (i.e., size, amount) can be achieved using a deep convolutional neural network that predicts wound tissue types. The proposed objective measurement by machine learning model’s performance is reported in terms of mean intersection over union (mIOU) between model prediction and the ground truth truth labels. Finally, to compare the performance of the model wound tissue identification by a cohort of wound care clinicians.

Methods:

A dataset of 58 anonymized wound images of various types of chronic wounds from Swift Medical’s Wound Database was used to conduct the inter-rater and intra-rater agreement study. The dataset was split into 3 subsets, with 50% overlap between subsets to measure intra-rater agreement. Four different tissue types (epithelial, granulation, slough and eschar) within the wound bed were independently labeled by the 5 wound clinicians using a browser-based image annotation tool. Each subset was labeled at one-week intervals. Inter-rater and intra rater agreement ws computed. Next, two separate deep convolutional neural networks architectures were developed for wound segmentation and tissue segmentation and are used in sequence in the proposed workflow. These models were trained using 465,187 wound image-label pairs and 17,000 image-label pairs respectively. This is by far the largest and most diverse reported dataset of labeled wound images used for training deep learning models for wound and wound tissue segmentation. This allows our models to be robust, unbiased towards skin tones and generalize well to unseen data. The deep learning model architectures were designed to be fast and nimble to allow them to run in near real time on mobile devices.

Results:

We observed considerable variability when a cohort of wound clinicians were tasked to label the different tissue types within the wound using a browser based image annotation tool. We report poor to moderate inter-rater agreement in identifying tissue types in chronic wound images. A very poor Kripendorff alpha value of 0.014 for inter-rater variability when identifying epithelization has been observed, while granulation is most consistently identified by the clinicians. The intra-rater ICC(3,1) (Intra-Class Correlation) however indicates raters are relatively consistent when labeling the same image multiple times over a period of time. Our deep learning models achieved a mean intersection over union (mIOU) of 0.8644 and 0.7192 for wound and tissue segmentation respectively. A cohort of wound clinicians, by consensus, rated 91% of the tissue segmentation results to be between fair and good in terms of tissue identification and segmentation quality.

Conclusions:

Our inter-rater agreement study validates that clinicians may exhibit considerable variability when identifying and visually estimating tissue proportion within the wound bed. The proposed deep learning model provides objective tissue identification and measurements to assist clinicians in documenting the wound more accurately. Our solution works on off-the-shelf mobile devices, and was trained with the largest and most diverse chronic wound dataset ever reported and leading to a robust model when deployed. The proposed solution brings us a step closer to more accurate wound documentation and may lead to improved healing outcomes when deployed at scale.


 Citation

Please cite as:

Ramachandram D, Ramirez-GarciaLuna J, Fraser RD, Martinez-Jimenez M, Arriaga-Caballero J, Allport J

Fully Automated Wound Tissue Segmentation Using Deep Learning on Mobile Devices: Cohort Study

JMIR Mhealth Uhealth 2022;10(4):e36977

DOI: 10.2196/36977

PMID: 35451982

PMCID: 9077502

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.