Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Wednesday, July 01, 2020 at 8:00 PM to 10:00 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Accepted for/Published in: JMIR Medical Education

Date Submitted: Nov 5, 2022
Date Accepted: Mar 31, 2023
Date Submitted to PubMed: Mar 31, 2023

The final, peer-reviewed published version of this preprint can be found here:

Scoring Single-Response Multiple-Choice Items: Scoping Review and Comparison of Different Scoring Methods

Kanzow AF, Schmidt D, Kanzow P

Scoring Single-Response Multiple-Choice Items: Scoping Review and Comparison of Different Scoring Methods

JMIR Med Educ 2023;9:e44084

DOI: 10.2196/44084

PMID: 37001510

PMCID: 10238964

Scoring Single-Response Multiple-Choice Items – Quite Simple?! A Scoping Review and Comparison of Different Scoring Methods

  • Amelie Friederike Kanzow; 
  • Dennis Schmidt; 
  • Philipp Kanzow

ABSTRACT

Background:

Single-choice items (eg, best-answer items, alternate-choice items, single true-false items) are one type of multiple-choice items and have been used in examinations for over 100 years. At the end of every examination, the examinees' responses have to be analyzed and scored in order to derive with an information about examinees' true knowledge.

Objective:

The aim of this paper is to compile scoring methods for individual single-choice items described in the literature. Furthermore, the metric expected chance score and the relation between examinees' true knowledge and expected scoring results (averaged percentage score) are analyzed. Furthermore, implications for potential pass marks to be used in examinations to test examinees for a predefined level of true knowledge are derived.

Methods:

Scoring methods for individual single-choice items including were extracted from various databases (ERIC, PsycInfo, Embase via Ovid, MEDLINE via PubMed) in September 2020. Eligible sources reported on scoring methods for individual single-choice items in written examinations including but not limited to medical education. Separately for items with n = 2 answer options (eg, alternate-choice items, single true-false items) and best-answer items with n = 5 answer options (eg, Type A items) and for each identified scoring method, the metric expected chance score and the expected scoring results as a function of examinees' true knowledge using fictitious examinations with 100 single-choice items were calculated.

Results:

A total of 21 different scoring methods were identified from the 258 included sources, with varying consideration of correctly marked, omitted, and incorrectly marked items. Resulting credit varied between -3 and +1 credit points per item. For items with n = 2 answer options, expected chance scores from random guessing ranged between -1 and +0.75 credit points. For items with n = 5 answer options, expected chance scores ranged between -2.2 and +0.84 credit points. All scoring methods showed a linear relation between examinees' true knowledge and the expected scoring results. Depending on the scoring method used, examination results differed considerably: Expected scoring results from examinees with 50% true knowledge ranged between 0.0% (95% CI: 0% to 0%) and 87.5% (95% CI: 81.0% to 94.0%) for items with n = 2 and between -60.0% (95% CI: -60% to -60%) and 92.0% (95% CI: 86.7% to 97.3%) for items with n = 5.

Conclusions:

In examinations with single-choice items, the scoring result is not always equivalent to examinees' true knowledge. When interpreting examination scores and setting pass marks, the number of answer options per item must usually be taken into account in addition to the scoring method used.


 Citation

Please cite as:

Kanzow AF, Schmidt D, Kanzow P

Scoring Single-Response Multiple-Choice Items: Scoping Review and Comparison of Different Scoring Methods

JMIR Med Educ 2023;9:e44084

DOI: 10.2196/44084

PMID: 37001510

PMCID: 10238964

Download PDF


Request queued. Please wait while the file is being generated. It may take some time.

© The authors. All rights reserved. This is a privileged document currently under peer-review/community review (or an accepted/rejected manuscript). Authors have provided JMIR Publications with an exclusive license to publish this preprint on it's website for review and ahead-of-print citation purposes only. While the final peer-reviewed paper may be licensed under a cc-by license on publication, at this stage authors and publisher expressively prohibit redistribution of this draft paper other than for review purposes.