par ncbi
Open source · 259k downloads · 60 likes
Le modèle MedCPT Query Encoder est spécialisé dans la génération d'embeddings pour des textes biomédicaux courts, comme des requêtes ou des phrases, afin de faciliter la recherche sémantique par similarité. Entraîné sur un vaste ensemble de 255 millions de paires requête-article issues des logs de recherche PubMed, il excelle dans des tâches de recherche d'information biomédicale sans entraînement préalable, surpassant d'autres modèles sur plusieurs benchmarks. Ses principales applications incluent la recherche d'articles pertinents à partir d'une requête, le regroupement de requêtes similaires ou d'articles, ainsi que la comparaison directe entre textes courts et longs grâce à une représentation vectorielle commune. Ce qui le distingue est sa capacité à capturer des nuances spécifiques au domaine biomédical, optimisant ainsi la pertinence des résultats dans ce champ exigeant. Développé par le NCBI, il s'appuie sur des données réelles et des besoins concrets en recherche médicale.
MedCPT generates embeddings of biomedical texts that can be used for semantic search (dense retrieval). The model contains two encoders:
This repo contains the MedCPT Query Encoder.
MedCPT has been pre-trained by an unprecedented scale of 255M query-article pairs from PubMed search logs, and has been shown to achieve state-of-the-art performance on several zero-shot biomedical IR datasets. In general, there are three use cases:
For more details, please check out our paper (Bioinformatics, 2023). Please note that the released version is slightly different from the version reported in the paper.
import torch
from transformers import AutoTokenizer, AutoModel
model = AutoModel.from_pretrained("ncbi/MedCPT-Query-Encoder")
tokenizer = AutoTokenizer.from_pretrained("ncbi/MedCPT-Query-Encoder")
queries = [
"diabetes treatment",
"How to treat diabetes?",
"A 45-year-old man presents with increased thirst and frequent urination over the past 3 months.",
]
with torch.no_grad():
# tokenize the queries
encoded = tokenizer(
queries,
truncation=True,
padding=True,
return_tensors='pt',
max_length=64,
)
# encode the queries (use the [CLS] last hidden states as the representations)
embeds = model(**encoded).last_hidden_state[:, 0, :]
print(embeds)
print(embeds.size())
The output will be:
tensor([[ 0.0413, 0.0084, -0.0491, ..., -0.4963, -0.3830, -0.3593],
[ 0.0801, 0.1193, -0.0905, ..., -0.5380, -0.5059, -0.2944],
[-0.3412, 0.1521, -0.0946, ..., 0.0952, 0.1660, -0.0902]])
torch.Size([3, 768])
These embeddings are also in the same space as those generated by the MedCPT article encoder.
We have provided the embeddings of all PubMed articles generated by the MedCPT article encoder at https://ftp.ncbi.nlm.nih.gov/pub/lu/MedCPT/pubmed_embeddings/. You can simply download these embeddings to search PubMed with your query.
This work was supported by the Intramural Research Programs of the National Institutes of Health, National Library of Medicine.
This tool shows the results of research conducted in the Computational Biology Branch, NCBI/NLM. The information produced on this website is not intended for direct diagnostic use or medical decision-making without review and oversight by a clinical professional. Individuals should not change their health behavior solely on the basis of information produced on this website. NIH does not independently verify the validity or utility of the information produced by this tool. If you have questions about the information produced on this website, please see a health care professional. More information about NCBI's disclaimer policy is available.
If you find this repo helpful, please cite MedCPT by:
@article{jin2023medcpt,
title={MedCPT: Contrastive Pre-trained Transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval},
author={Jin, Qiao and Kim, Won and Chen, Qingyu and Comeau, Donald C and Yeganova, Lana and Wilbur, W John and Lu, Zhiyong},
journal={Bioinformatics},
volume={39},
number={11},
pages={btad651},
year={2023},
publisher={Oxford University Press}
}