RoBERTa-base 125M (fine-tuned) | 27.9 | RoBERTa: A Robustly Optimized BERT Pretraining Approach | |
GPT-3 175B (5-shot) | 43.9 | Language Models are Few-Shot Learners | |
Flan-T5-Base 250M (CoT) | 33.7 | Scaling Instruction-Finetuned Language Models | |
Claude Instant 1.1 (5-shot) | 73.4 | Model Card and Evaluations for Claude Models | - |