Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Nursing

Date Submitted: Mar 6, 2024
Open Peer Review Period: Apr 4, 2024 - May 30, 2024
Date Accepted: Sep 23, 2024
(closed for review but you can still tweet)

The final, peer-reviewed published version of this preprint can be found here:

Unobtrusive Nighttime Movement Monitoring to Support Nursing Home Continence Care: Algorithm Development and Validation Study

Strauven H, Wang C, Hallez H, Vanden Abeele V, Vanrumste B

Unobtrusive Nighttime Movement Monitoring to Support Nursing Home Continence Care: Algorithm Development and Validation Study

JMIR Nursing 2024;7:e58094

DOI: 10.2196/58094

PMID: 39718558

PMCID: 11729779

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

Unobtrusive Nighttime Movement Monitoring to Support Nursing Home Continence Care: Machine Learning Study

  • Hannelore Strauven; 
  • Chunzhuo Wang; 
  • Hans Hallez; 
  • Vero Vanden Abeele; 
  • Bart Vanrumste

ABSTRACT

Background:

The rising prevalence of urinary incontinence (UI) among older adults, particularly those living in nursing homes (NHs), underscores the need for innovative continence care solutions. The implementation of an unobtrusive sensor system may support nighttime monitoring of NH residents' movements and, more specifically, the agitation possibly associated with voiding events.

Objective:

This study explores the application of an unobtrusive sensor system to monitor nighttime movement, integrated into a care bed with accelerometer sensors connected to a pressure redistributing care mattress.

Methods:

Six participants followed a seven-step protocol. The obtained dataset was segmented into 20s windows with a 50% overlap. Each window was labelled with one of the four chosen activity classes: in bed, agitation, turn and out of bed. A total of 1416 features were selected and analyzed with an XGBoost algorithm. At last, the model was validated using ‘leave one subject out cross-validation’ (LOSOCV).

Results:

The trained model attained a trustworthy overall F1-score of 81.51% for all classes, and, more specifically, an F1-score of 83.87% for the class 'Agitation'.

Conclusions:

The results from this study provide promising insights in unobtrusive nighttime movement monitoring. The study underscores the potential to enhance the quality of care for NH residents, via a machine learning model based on data from accelerometers connected to a viscoelastic care mattress, thereby driving progress in the field of continence care and AI-supported healthcare for older adults. Clinical Trial: Ethical approval to conduct the research was obtained from the KU Leuven Social and Societal Ethics Committee with protocol number G-2020-2214.


 Citation

Please cite as:

Strauven H, Wang C, Hallez H, Vanden Abeele V, Vanrumste B

Unobtrusive Nighttime Movement Monitoring to Support Nursing Home Continence Care: Algorithm Development and Validation Study

JMIR Nursing 2024;7:e58094

DOI: 10.2196/58094

PMID: 39718558

PMCID: 11729779

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.