Scoring Single-Response Multiple-Choice Items: Scoping Review and Comparison of Different Scoring Methods
- PMID: 37001510
- PMCID: PMC10238964
- DOI: 10.2196/44084
Scoring Single-Response Multiple-Choice Items: Scoping Review and Comparison of Different Scoring Methods
Abstract
Background: Single-choice items (eg, best-answer items, alternate-choice items, single true-false items) are 1 type of multiple-choice items and have been used in examinations for over 100 years. At the end of every examination, the examinees' responses have to be analyzed and scored to derive information about examinees' true knowledge.
Objective: The aim of this paper is to compile scoring methods for individual single-choice items described in the literature. Furthermore, the metric expected chance score and the relation between examinees' true knowledge and expected scoring results (averaged percentage score) are analyzed. Besides, implications for potential pass marks to be used in examinations to test examinees for a predefined level of true knowledge are derived.
Methods: Scoring methods for individual single-choice items were extracted from various databases (ERIC, PsycInfo, Embase via Ovid, MEDLINE via PubMed) in September 2020. Eligible sources reported on scoring methods for individual single-choice items in written examinations including but not limited to medical education. Separately for items with n=2 answer options (eg, alternate-choice items, single true-false items) and best-answer items with n=5 answer options (eg, Type A items) and for each identified scoring method, the metric expected chance score and the expected scoring results as a function of examinees' true knowledge using fictitious examinations with 100 single-choice items were calculated.
Results: A total of 21 different scoring methods were identified from the 258 included sources, with varying consideration of correctly marked, omitted, and incorrectly marked items. Resulting credit varied between -3 and +1 credit points per item. For items with n=2 answer options, expected chance scores from random guessing ranged between -1 and +0.75 credit points. For items with n=5 answer options, expected chance scores ranged between -2.2 and +0.84 credit points. All scoring methods showed a linear relation between examinees' true knowledge and the expected scoring results. Depending on the scoring method used, examination results differed considerably: Expected scoring results from examinees with 50% true knowledge ranged between 0.0% (95% CI 0% to 0%) and 87.5% (95% CI 81.0% to 94.0%) for items with n=2 and between -60.0% (95% CI -60% to -60%) and 92.0% (95% CI 86.7% to 97.3%) for items with n=5.
Conclusions: In examinations with single-choice items, the scoring result is not always equivalent to examinees' true knowledge. When interpreting examination scores and setting pass marks, the number of answer options per item must usually be taken into account in addition to the scoring method used.
Keywords: Type A; alternate-choice; best-answer; education; education system; educational assessment; educational measurement; examination; multiple choice; results; scoping review; scoring; scoring system; single choice; single response; test; testing; true-false; true/false.
©Amelie Friederike Kanzow, Dennis Schmidt, Philipp Kanzow. Originally published in JMIR Medical Education (https://mededu.jmir.org), 19.05.2023.
Conflict of interest statement
Conflicts of Interest: None declared.
Figures



Similar articles
-
Use of Multiple-Choice Items in Summative Examinations: Questionnaire Survey Among German Undergraduate Dental Training Programs.JMIR Med Educ. 2024 Jun 27;10:e58126. doi: 10.2196/58126. JMIR Med Educ. 2024. PMID: 38952022 Free PMC article.
-
Use of Multiple-Select Multiple-Choice Items in a Dental Undergraduate Curriculum: Retrospective Study Involving the Application of Different Scoring Methods.JMIR Med Educ. 2023 Mar 27;9:e43792. doi: 10.2196/43792. JMIR Med Educ. 2023. PMID: 36841970 Free PMC article.
-
Multiple true-false items: a comparison of scoring algorithms.Adv Health Sci Educ Theory Pract. 2018 Aug;23(3):455-463. doi: 10.1007/s10459-017-9805-y. Epub 2017 Nov 30. Adv Health Sci Educ Theory Pract. 2018. PMID: 29189963
-
The psychometric properties of five scoring methods applied to the script concordance test.Acad Med. 2005 Apr;80(4):395-9. doi: 10.1097/00001888-200504000-00019. Acad Med. 2005. PMID: 15793026 Clinical Trial.
-
Threats to validity in the use and interpretation of script concordance test scores.Med Educ. 2013 Dec;47(12):1175-83. doi: 10.1111/medu.12283. Med Educ. 2013. PMID: 24206151 Review.
Cited by
-
Item analysis: the impact of distractor efficiency on the difficulty index and discrimination power of multiple-choice items.BMC Med Educ. 2024 Apr 24;24(1):445. doi: 10.1186/s12909-024-05433-y. BMC Med Educ. 2024. PMID: 38658912 Free PMC article.
-
Assessing the Climate Readiness of Physician Education Leaders in Graduate Medical Education.J Patient Cent Res Rev. 2024 Oct 15;11(3):231-236. doi: 10.17294/2330-0698.2112. eCollection 2024 Fall. J Patient Cent Res Rev. 2024. PMID: 39439540 Free PMC article.
-
Use of Multiple-Choice Items in Summative Examinations: Questionnaire Survey Among German Undergraduate Dental Training Programs.JMIR Med Educ. 2024 Jun 27;10:e58126. doi: 10.2196/58126. JMIR Med Educ. 2024. PMID: 38952022 Free PMC article.
References
-
- Krebs R. Prüfen mit Multiple Choice: Kompetent planen, entwickeln, durchführen und auswerten [Testing with Multiple Choice: Plan, Develop, Implement, and Evaluate Competently] Bern, Switzerland: Hogrefe; 2019.
-
- Ebel RL. Proposed solutions to two problems of test construction. J Educ Meas. 1982 Dec;19(4):267–278. doi: 10.1111/j.1745-3984.1982.tb00133.x. - DOI
-
- Kelly FJ. The Kansas silent reading test. J Educ Psychol. 1916 Feb;7(2):63–80. doi: 10.1037/h0073542. - DOI
-
- McCall WA. A new kind of school examination. J Educ Res. 1920;1(1):33–46. doi: 10.1080/00220671.1920.10879021. - DOI
-
- Ruch GM, Stoddard GD. Comparative reliabilities of five types of objective examinations. J Educ Psychol. 1925 Feb;16(2):89–103. doi: 10.1037/h0072894. - DOI
Publication types
LinkOut - more resources
Full Text Sources