HyperAI超神経

Language Modelling On Lambada

評価指標

Accuracy

評価結果

このベンチマークにおける各モデルのパフォーマンス結果

モデル名
Accuracy
Paper TitleRepository
OPT-175B (50% Sparsity)0.02SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
GLM-XXLarge (bidirectional)72.35GLM: General Language Model Pretraining with Autoregressive Blank Infilling
GPT-3 175B (Few-Shot)86.4Language Models are Few-Shot Learners
test0.01Test-Time Training with Self-Supervision for Generalization under Distribution Shifts
Megatron-Turing NLG 530B (Few-Shot)Megatron-Turing NLG 530B (Few-Shot)Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model-
PaLM 2-M (one-shot)83.7PaLM 2 Technical Report
GPT-3 13B (Zero-Shot)72.5Language Models are Few-Shot Learners
Gated-Attention Reader (+ features)49.0Broad Context Language Modeling as Reading Comprehension-
GPT-2 1.5B (Zero Shot)63.24Language Models are Unsupervised Multitask Learners-
GPT-3 2.7B (Zero-Shot)67.1Language Models are Few-Shot Learners
Pythia 12B(Zero-Shot)-Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
PaLM 2-L (one-shot)86.9PaLM 2 Technical Report
LLaMA-30B+CFG (zero-shot)83.9Stay on topic with Classifier-Free Guidance-
Universal Transformer (w/ dynamic halting)56.25Universal Transformers
SparseGPT (175B, 2:4 Sparsity)79.47SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
SparseGPT (175B, 50% Sparsity)76.51SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
PaLM-540B (Zero-Shot)77.9PaLM: Scaling Language Modeling with Pathways
Pythia 6.9B (0-shot)67.28Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
GLaM 62B/64E (One-Shot)80.9GLaM: Efficient Scaling of Language Models with Mixture-of-Experts-
Residual Shuffle-Exchange network54.34Residual Shuffle-Exchange Networks for Fast Processing of Long Sequences
0 of 37 row(s) selected.