Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR mHealth and uHealth

Date Submitted: Aug 4, 2021
Date Accepted: Feb 2, 2022
Date Submitted to PubMed: Apr 18, 2022

The final, peer-reviewed published version of this preprint can be found here:

A Novel Method for Evaluating Mobile Apps (App Rating Inventory): Development Study

Mackey R, Gleason A, Ciulla R

A Novel Method for Evaluating Mobile Apps (App Rating Inventory): Development Study

JMIR Mhealth Uhealth 2022;10(4):e32643

DOI: 10.2196/32643

PMID: 35436227

PMCID: 9055478

Warning: This is an author submission that is not peer-reviewed or edited. Preprints - unless they show as "accepted" - should not be relied on to guide clinical practice or health-related behavior and should not be reported in news media as established information.

The App Rating Inventory: A Novel Method for Evaluating Mobile Apps

  • Rachel Mackey; 
  • Ann Gleason; 
  • Robert Ciulla

ABSTRACT

Background:

A prodigious number of mobile health apps have flooded the market. The lack of guidelines for identifying high-quality apps from the overwhelming number of available apps creates confusion forestalling clinical adoption.

Objective:

The Defense Health Agency’s (DHA) Connected Health Branch developed the app rating inventory (ARI), an objective rating system with capability for broad application across condition areas.

Methods:

During the development of the ARI, three rounds of testing were conducted to enhance the tool’s performance, reduce redundancy, validate the ARI’s broad application, and assess potential subjectivity.

Results:

The ARI is a 28-item, three-criterion tool. The evidence criterion contains six items, and the content and customizability criterion each contain eleven items. Scoring is based on a simple binary system: either the app contains the feature or it does not. The 28 items are weighted equally; no one item is considered more (or less) important than any other. Each rated app receives four scores: a score for evidence, content, and customizability, and a total score (the sum of the three categories.). Higher scores indicate that the app obtained a positive score on more items than a similar app with a lower score. The evidence, content and customizability scores allow a clinician to make focused decisions when selecting an app for clinical use.

Conclusions:

Using a two-phased process (market research followed by ratings), the ARI is able to evaluate apps for evidence, content and customizability. Scoring systems provide guidance; they filter down hundreds of apps in a disease category to a handful for consideration. Indeed, apps are not new medicine; in many cases, they are a novel delivery system for proven interventions.


 Citation

Please cite as:

Mackey R, Gleason A, Ciulla R

A Novel Method for Evaluating Mobile Apps (App Rating Inventory): Development Study

JMIR Mhealth Uhealth 2022;10(4):e32643

DOI: 10.2196/32643

PMID: 35436227

PMCID: 9055478

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.