Multiple Choice Question Answering Mcqa On 25
Metriken
Accuracy
Ergebnisse
Leistungsergebnisse verschiedener Modelle zu diesem Benchmark
Vergleichstabelle
Modellname | Accuracy |
---|---|
towards-expert-level-medical-question | 92.3 |
towards-expert-level-medical-question | 95.2 |
biomedgpt-open-multimodal-generative-pre | 51.1 |
towards-expert-level-medical-question | 93.4 |
llama-2-open-foundation-and-fine-tuned-chat | 43.38 |
llama-2-open-foundation-and-fine-tuned-chat | 40.07 |