Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Comparative Study
. 2010 Oct 12:10:58.
doi: 10.1186/1472-6947-10-58.

Boolean versus ranked querying for biomedical systematic reviews

Affiliations
Comparative Study

Boolean versus ranked querying for biomedical systematic reviews

Sarvnaz Karimi et al. BMC Med Inform Decis Mak. .

Abstract

Background: The process of constructing a systematic review, a document that compiles the published evidence pertaining to a specified medical topic, is intensely time-consuming, often taking a team of researchers over a year, with the identification of relevant published research comprising a substantial portion of the effort. The standard paradigm for this information-seeking task is to use Boolean search; however, this leaves the user(s) the requirement of examining every returned result. Further, our experience is that effective Boolean queries for this specific task are extremely difficult to formulate and typically require multiple iterations of refinement before being finalized.

Methods: We explore the effectiveness of using ranked retrieval as compared to Boolean querying for the purpose of constructing a systematic review. We conduct a series of experiments involving ranked retrieval, using queries defined methodologically, in an effort to understand the practicalities of incorporating ranked retrieval into the systematic search task.

Results: Our results show that ranked retrieval by itself is not viable for this search task requiring high recall. However, we describe a refinement of the standard Boolean search process and show that ranking within a Boolean result set can improve the overall search performance by providing early indication of the quality of the results, thereby speeding up the iterative query-refinement process.

Conclusions: Outcomes of experiments suggest that an interactive query-development process using a hybrid ranked and Boolean retrieval system has the potential for significant time-savings over the current search process in the systematic reviewing.

PubMed Disclaimer

Figures

Figure 1
Figure 1
A sample ranked query (TRC) based on an AHRQ systematic review (Drug Dataset). Note the headings only show where the query words are sourced and do not represent any structure to the query. The query is treated as a "bag-of-words", that is, the entire text is entered as a query without imposing any structure or word order.
Figure 2
Figure 2
Tolerance versus recall (left), and rank (right). Lines show the relationship for the best (query 14), worst (query 1), and average over 15 TRC queries, when retrieving from the full MEDLINE collection. Evaluations are based on final included studies (Tier 2).
Figure 3
Figure 3
Average recall versus precision using TRC ranked queries on MEDLINE.
Figure 4
Figure 4
Recall of Boolean queries after re-running on Ovid MEDLINE (1950-Nov week 3, 2008). Average recall for (a) original queries (upper) are 0.27 and 0.29 for Tier 1 and Tier 2, respectively, and for (b) simplified queries (lower) are 0.79 and 0.82.
Figure 5
Figure 5
Tolerance versus recall (left), and rank (right) for two queries (best and worst), and average over 15 TRC queries. Retrieval was over subsets of MEDLINE determined by the simplified Boolean queries. Evaluations are based on final included studies (Tier 2).
Figure 6
Figure 6
Average recall versus precision using TRC ranked queries on the subset of MEDLINE retrieved by their corresponding Boolean queries.
Figure 7
Figure 7
Recall of the Boolean queries in the Misc dataset after running the original (upper) and simplified (lower) Boolean queries on OVID MEDLINE.

References

    1. Haynes B, McKibbon KA, Wilczynski NL, Walter SD, Werre SR. Optimal search strategies for retrieving scientifically strong studies of treatment from MEDLINE: analytical survey. British Medical Journal. 2005;330(7501):1179–1182. doi: 10.1136/bmj.38446.498542.8F. - DOI - PMC - PubMed
    1. Zhang L, Ajiferuke I, Sampson M. Optimizing search strategies to identify randomized con-trolled trials in MEDLINE. BMC Medical Research Methodology. 2006;6(23) - PMC - PubMed
    1. The Cochrane Collaboration. Cochrane Handbook for Systematic Reviews of Interventions, Version 5.0.0. 2008. http://www.cochrane.org/resources/handbook/
    1. Klassen TP, Jadad AR, Moher D. Guides for reading and interpreting systematic reviews: I. Getting started. Archives of pediatrics and adolescent medicine. 1998;152(7):700–704. - PubMed
    1. Dickersin K, Scherer R, Lefebvre C. Systematic Reviews: Identifying relevant studies for systematic reviews. British Medical Journal. 1994;309(6964):1286–1291. - PMC - PubMed

Publication types

MeSH terms

LinkOut - more resources