Question Answering On Blurb
Métriques
Accuracy
Résultats
Résultats de performance de divers modèles sur ce benchmark
Nom du modèle | Accuracy | Paper Title | Repository |
---|---|---|---|
BioLinkBERT (base) | 80.81 | LinkBERT: Pretraining Language Models with Document Links | |
BioLinkBERT (large) | 83.5 | LinkBERT: Pretraining Language Models with Document Links | |
PubMedBERT (uncased; abstracts) | 71.7 | Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing | |
GPT-4 | 80.56 | Evaluation of large language model performance on the Biomedical Language Understanding and Reasoning Benchmark | - |
0 of 4 row(s) selected.