Question Answering On Uniprotqa
Metrics
BLEU-2
BLEU-4
MEATOR
ROUGE-1
ROUGE-2
ROUGE-L
Results
Performance results of various models on this benchmark
Comparison Table
Model Name | BLEU-2 | BLEU-4 | MEATOR | ROUGE-1 | ROUGE-2 | ROUGE-L |
---|---|---|---|---|---|---|
llama-2-open-foundation-and-fine-tuned-chat | 0.019 | 0.002 | 0.052 | 0.103 | 0.060 | 0.009 |
biomedgpt-open-multimodal-generative-pre | 0.571 | 0.535 | 0.754 | 0.743 | 0.759 | 0.622 |