MedCPT: Contrastive Pre-trained Transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval
- PMID: 37930897
- PMCID: PMC10627406
- DOI: 10.1093/bioinformatics/btad651
MedCPT: Contrastive Pre-trained Transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval
Abstract
Motivation: Information retrieval (IR) is essential in biomedical knowledge acquisition and clinical decision support. While recent progress has shown that language model encoders perform better semantic retrieval, training such models requires abundant query-article annotations that are difficult to obtain in biomedicine. As a result, most biomedical IR systems only conduct lexical matching. In response, we introduce MedCPT, a first-of-its-kind Contrastively Pre-trained Transformer model for zero-shot semantic IR in biomedicine.
Results: To train MedCPT, we collected an unprecedented scale of 255 million user click logs from PubMed. With such data, we use contrastive learning to train a pair of closely integrated retriever and re-ranker. Experimental results show that MedCPT sets new state-of-the-art performance on six biomedical IR tasks, outperforming various baselines including much larger models, such as GPT-3-sized cpt-text-XL. In addition, MedCPT also generates better biomedical article and sentence representations for semantic evaluations. As such, MedCPT can be readily applied to various real-world biomedical IR tasks.
Availability and implementation: The MedCPT code and model are available at https://github.com/ncbi/MedCPT.
Published by Oxford University Press 2023.
Conflict of interest statement
None declared.
Figures
Update of
-
MedCPT: Contrastive Pre-trained Transformers with Large-scale PubMed Search Logs for Zero-shot Biomedical Information Retrieval.ArXiv [Preprint]. 2023 Oct 4:arXiv:2307.00589v2. ArXiv. 2023. Update in: Bioinformatics. 2023 Nov 1;39(11):btad651. doi: 10.1093/bioinformatics/btad651. PMID: 41031073 Free PMC article. Updated. Preprint.
References
-
- Brown T, Mann B, Ryder N. et al. Language models are few-shot learners. Adv Neural Inf Process Syst 2020;33:1877–901.
-
- Chen Q, Peng Y, Lu Z. BioSentVec: creating sentence embeddings for biomedical texts. In: 2019 IEEE International Conference on Healthcare Informatics (ICHI). Xian, China: IEEE, 2019, 1–5.
-
- Cohan A, Feldman S, Beltagy I. et al. SPECTER: document-level representation learning using citation-informed transformers. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Online, 2020, 2270–82.
