Putting double marking to the test: a framework to assess if it is worth the trouble
- PMID: 15733166
- DOI: 10.1111/j.1365-2929.2005.02093.x
Putting double marking to the test: a framework to assess if it is worth the trouble
Abstract
Background: It is a challenge to assign a mark that accurately measures the quality of students' work in essay-type assessments that require an element of judgement and fairness by the markers. Double marking such assessments has been seen as a way of improving the reliability of the mark. The analysis approach often taken is to look for absolute agreement between markers instead of looking at all aspects of reliability.
Aim: To develop an analytic process that will examine the components and meanings of reliability calculations that can be used for assessing the value of double marking a piece of work.
Methods: An undergraduate case study assessment in General Practice was used as an illustration. Datasets of double marking were collected retrospectively for 1999-2000, and prospectively for 2002-03. An assessment of intermarker agreement and its effect on the reliability of the final mark for students was made, using methods dependent on the type of data collected and Generalisability Theory.
Results and conclusions: The data were used to illustrate how to interpret the results of Bland and Altman plots, anova tables and Cohen's kappa calculations. Generalisability Theory was used to show that, while there was reasonable agreement between markers, the reliability of the mark for the student was still only moderate, probably due to unexplained variability elsewhere in the process. Possible reasons for this variability are discussed. A flowchart of the decisions and actions needed to judge whether a piece of work should be double marked has been constructed.
Similar articles
-
Reliability: on the reproducibility of assessment data.Med Educ. 2004 Sep;38(9):1006-12. doi: 10.1111/j.1365-2929.2004.01932.x. Med Educ. 2004. PMID: 15327684
-
Feasibility and reliability of an in-training assessment programme in an undergraduate clerkship.Med Educ. 2004 Dec;38(12):1270-7. doi: 10.1111/j.1365-2929.2004.02019.x. Med Educ. 2004. PMID: 15566538
-
The effectiveness and reliability of peer-marking in first-year medical students.Med Educ. 2006 Oct;40(10):965-72. doi: 10.1111/j.1365-2929.2006.02565.x. Med Educ. 2006. PMID: 16987186 Clinical Trial.
-
Exploring students' perceptions on the use of significant event analysis, as part of a portfolio assessment process in general practice, as a tool for learning how to use reflection in learning.BMC Med Educ. 2007 Mar 30;7:5. doi: 10.1186/1472-6920-7-5. BMC Med Educ. 2007. PMID: 17397544 Free PMC article.
-
Peer assessment of professional competence.Med Educ. 2005 Jul;39(7):713-22. doi: 10.1111/j.1365-2929.2005.02193.x. Med Educ. 2005. PMID: 15960792
Cited by
-
Learning to mark: a qualitative study of the experiences and concerns of medical markers.BMC Med Educ. 2006 Apr 25;6:25. doi: 10.1186/1472-6920-6-25. BMC Med Educ. 2006. PMID: 16638150 Free PMC article.
-
Use of cost-effectiveness analysis to compare the efficiency of study identification methods in systematic reviews.Syst Rev. 2016 Aug 17;5(1):140. doi: 10.1186/s13643-016-0315-4. Syst Rev. 2016. PMID: 27535658 Free PMC article.
MeSH terms
LinkOut - more resources
Full Text Sources