HyperAI

Common Sense Reasoning On Arc Easy

Metrics

Accuracy

Results

Performance results of various models on this benchmark

Model Name
Accuracy
Paper TitleRepository
LLaMA 13B (0-shot)74.8LLaMA: Open and Efficient Foundation Language Models
GLaM 64B/64E (0-shot)68.0GLaM: Efficient Scaling of Language Models with Mixture-of-Experts-
GPT-3 175B (1 shot)71.2Language Models are Few-Shot Learners
ST-MoE-32B 269B (fine-tuned)95.2ST-MoE: Designing Stable and Transferable Sparse Expert Models-
SparseGPT 175B (50% sparsity)69.65SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
Mamba-2.8B (0-shot)69.7Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Camelidae-8×34B86.2Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General Tasks
ST-MoE-L 4.1B (fine-tuned)75.4ST-MoE: Designing Stable and Transferable Sparse Expert Models-
LLaMA 7B (0-shot)72.8LLaMA: Open and Efficient Foundation Language Models
UL2 20B (0-shot)32.2UL2: Unifying Language Learning Paradigms
SparseGPT (175B, 4:8 Sparsity)68.35SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
Pythia 12B (0-shot)70.2Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
LLaMA 13B + CFG (0-shot)79.1Stay on topic with Classifier-Free Guidance-
Mistral 7B (0-shot)80.5Mixtral of Experts
LLaMA 65B + CFG (0-shot)84.2Stay on topic with Classifier-Free Guidance-
LLaMA 33B (0-shot)80.0LLaMA: Open and Efficient Foundation Language Models
LLaMA-2 13B + MixLoRA83.5MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts
LLaMA-2 7B + MixLoRA77.7MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of Experts
LLaMA 3 8B+MoSLoRA (fine-tuned)90.5Mixture-of-Subspaces in Low-Rank Adaptation
OPT-175B71.04SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
0 of 47 row(s) selected.