Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Randomized Controlled Trial
. 2013 Jan 17;15(1):e8.
doi: 10.2196/jmir.2055.

Human computation as a new method for evidence-based knowledge transfer in Web-based guideline development groups: proof of concept randomized controlled trial

Affiliations
Randomized Controlled Trial

Human computation as a new method for evidence-based knowledge transfer in Web-based guideline development groups: proof of concept randomized controlled trial

Annemie Heselmans et al. J Med Internet Res. .

Abstract

Background: Guideline developers use different consensus methods to develop evidence-based clinical practice guidelines. Previous research suggests that existing guideline development techniques are subject to methodological problems and are logistically demanding. Guideline developers welcome new methods that facilitate a methodologically sound decision-making process. Systems that aggregate knowledge while participants play a game are one class of human computation applications. Researchers have already proven that these games with a purpose are effective in building common sense knowledge databases.

Objective: We aimed to evaluate the feasibility of a new consensus method based on human computation techniques compared to an informal face-to-face consensus method.

Methods: We set up a randomized design to study 2 different methods for guideline development within a group of advanced students completing a master of nursing and obstetrics. Students who participated in the trial were enrolled in an evidence-based health care course. We compared the Web-based method of human-based computation (HC) with an informal face-to-face consensus method (IC). We used 4 clinical scenarios of lower back pain as the subject of the consensus process. These scenarios concerned the following topics: (1) medical imaging, (2) therapeutic options, (3) drugs use, and (4) sick leave. Outcomes were expressed as the amount of group (dis)agreement and the concordance of answers with clinical evidence. We estimated within-group and between-group effect sizes by calculating Cohen's d. We calculated within-group effect sizes as the absolute difference between the outcome value at round 3 and the baseline outcome value, divided by the pooled standard deviation. We calculated between-group effect sizes as the absolute difference between the mean change in outcome value across rounds in HC and the mean change in outcome value across rounds in IC, divided by the pooled standard deviation. We analyzed statistical significance of within-group changes between round 1 and round 3 using the Wilcoxon signed rank test. We assessed the differences between the HC and IC groups using Mann-Whitney U tests. We used a Bonferroni adjusted alpha level of .025 in all statistical tests. We performed a thematic analysis to explore participants' arguments during group discussion. Participants completed a satisfaction survey at the end of the consensus process.

Results: Of the 135 students completing a master of nursing and obstetrics, 120 participated in the experiment. We formed 8 HC groups (n=64) and 7 IC groups (n=56). The between-group comparison demonstrated that the human computation groups obtained a greater improvement in evidence scores compared to the IC groups, although the difference was not statistically significant. The between-group effect size was 0.56 (P=.30) for the medical imaging scenario, 0.07 (P=.97) for the therapeutic options scenario, and 0.89 (P=.11) for the drug use scenario. We found no significant differences in improvement in the degree of agreement between HC and IC groups. Between-group comparisons revealed that the HC groups showed greater improvement in degree of agreement for the medical imaging scenario (d=0.46, P=.37) and the drug use scenario (d=0.31, P=.59). Very few evidence arguments (6%) were quoted during informal group discussions.

Conclusions: Overall, the use of the IC method was appropriate as long as the evidence supported participants' beliefs or usual practice, or when the availability of the evidence was sparse. However, when some controversy about the evidence existed, the HC method outperformed the IC method. The findings of our study illustrate the importance of the choice of the consensus method in guideline development. Human computation could be an acceptable methodology for guideline development specifically for scenarios in which the evidence shows no resonance with participants' beliefs. Future research is needed to confirm the results of this study and to establish practical significance in a controlled setting of multidisciplinary guideline panels during real-life guideline development.

PubMed Disclaimer

Conflict of interest statement

Conflicts of Interest: The authors and evaluators are the developers of CPGame, a human-based computation application. There were no other conflicts of interest.

Figures

Figure 1
Figure 1
Screen capture of a multiple choice scenario in the CPGame application.
Figure 2
Figure 2
Screen capture of the information displayed on the CPGame when consensus was not reached within a team.
Figure 3
Figure 3
Screen capture of the final page in the CPGame application, showing the answers of all participants, the level of evidence, and the level of agreement between participants.
Figure 4
Figure 4
Flowchart showing participants in the trial.
Figure 5
Figure 5
Amount of (dis)agreement for the medical imaging scenario.
Figure 6
Figure 6
Amount of (dis)agreement for the therapeutic options scenario.
Figure 7
Figure 7
Amount of (dis)agreement for the drug use scenario.
Figure 8
Figure 8
Amount of (dis)agreement for the sick leave scenario.
Figure 9
Figure 9
Concordance with clinical evidence for the medical imaging scenario.
Figure 10
Figure 10
Concordance with clinical evidence for the therapeutic options scenario.
Figure 11
Figure 11
Concordance with clinical evidence for the drug use scenario.
Figure 12
Figure 12
Percentage of arguments in the different categories of the coding scheme.

Similar articles

References

    1. Commitee to Advise the Public Health Service on Clinical Practice Guidelines. Institute of Medicine . In: Clinical Practice Guidelines: Directions for a New Program. Field MJ, Lohr KN, editors. Washington, DC: National Academies Press; 1990. - PubMed
    1. Shekelle PG, Woolf SH, Eccles M, Grimshaw J. Clinical guidelines: developing guidelines. BMJ. 1999 Feb 27;318(7183):593–6. http://europepmc.org/abstract/MED/10037645 - PMC - PubMed
    1. Murphy MK, Black NA, Lamping DL, McKee CM, Sanderson CF, Askham J, Marteau T. Consensus development methods and their use in clinical guideline development. Health Technol Assess. 1998;2(3):i–iv, 1. http://www.hta.ac.uk/execsumm/summ203.htm - PubMed
    1. Gardner B, Davidson R, McAteer J, Michie S. A method for studying decision-making by guideline development groups. Implement Sci. 2009;4:48. doi: 10.1186/1748-5908-4-48. http://www.implementationscience.com/content/4//48 - DOI - PMC - PubMed
    1. Pagliari C, Grimshaw J, Eccles M. The potential influence of small group processes on guideline development. J Eval Clin Pract. 2001 May;7(2):165–73. - PubMed

Publication types