HyperAI초신경

Question Answering On Copa

평가 지표

Accuracy

평가 결과

이 벤치마크에서 각 모델의 성능 결과

비교 표
모델 이름Accuracy
winogrande-an-adversarial-winograd-schema90.6
efficient-language-modeling-with-sparse-all64
finetuned-language-models-are-zero-shot94
winogrande-an-adversarial-winograd-schema86.4
finetuned-language-models-are-zero-shot91
designing-effective-sparse-expert-models91
exploring-the-limits-of-transfer-learning92
language-models-are-few-shot-learners92
hungry-hungry-hippos-towards-language67
toward-efficient-language-model-pretraining99.4
hungry-hungry-hippos-towards-language51
exploring-the-limits-of-transfer-learning94.8
the-cot-collection-improving-zero-shot-and90.9
handling-multiword-expressions-in-causality69.9
ask-me-anything-a-simple-strategy-for77.0
knowledge-in-context-towards-knowledgeable85.30
toward-efficient-language-model-pretraining98.2
designing-effective-sparse-expert-models99.2
socialiqa-commonsense-reasoning-about-social80.8
palm-2-technical-report-190.0
alexatm-20b-few-shot-learning-using-a-large78.0
kelm-knowledge-enhanced-pre-trained-language78.0
bloomberggpt-a-large-language-model-for84
language-models-are-few-shot-learners73.0
palm-2-technical-report-189.0
handling-multiword-expressions-in-causality71.2
unifying-language-learning-paradigms85
palm-scaling-language-modeling-with-pathways-1100
unifying-language-learning-paradigms99
handling-multiword-expressions-in-causality70.2
hungry-hungry-hippos-towards-language67
ask-me-anything-a-simple-strategy-for58.2
efficient-language-modeling-with-sparse-all75
guess-the-instruction-making-language-models89.88
exploring-the-limits-of-transfer-learning83.4
language-models-are-few-shot-learners86
finetuned-language-models-are-zero-shot87
deberta-decoding-enhanced-bert-with98.4
socialiqa-commonsense-reasoning-about-social83.4
handling-multiword-expressions-in-causality58.8
efficient-language-modeling-with-sparse-all63
bloomberggpt-a-large-language-model-for86
hungry-hungry-hippos-towards-language77
winogrande-an-adversarial-winograd-schema76.4
bloomberggpt-a-large-language-model-for88
ask-me-anything-a-simple-strategy-for84.0
winogrande-an-adversarial-winograd-schema65.4
exploring-the-benefits-of-training-expert79.25
palm-2-technical-report-196.0
language-models-are-few-shot-learners91
winogrande-an-adversarial-winograd-schema84.4
deberta-decoding-enhanced-bert-with96.8
bloomberggpt-a-large-language-model-for86
hungry-hungry-hippos-towards-language81
n-grammer-augmenting-transformers-with-latent-160.0
efficient-language-modeling-with-sparse-all76
exploring-the-limits-of-transfer-learning71.2
language-models-are-few-shot-learners87
efficient-language-modeling-with-sparse-all79
handling-multiword-expressions-in-causality50