HyperAI

Language Modelling On The Pile

Metriken

Bits per byte

Ergebnisse

Leistungsergebnisse verschiedener Modelle zu diesem Benchmark

Modellname
Bits per byte
Paper TitleRepository
Test-Time Fine-Tuning with SIFT + Llama-3.2 (3B)0.557Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
Transformer 125M-Hungry Hungry Hippos: Towards Language Modeling with State Space Models
Jurassic-10.65GLM-130B: An Open Bilingual Pre-trained Model
Llama-3.2 3B0.640Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
Phi-3 3.8B0.679Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
GPT-3 Davinci 175B (pre-trained)0.7177The Pile: An 800GB Dataset of Diverse Text for Language Modeling
Gemma-2 9B0.670Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
Gemma-2 27B0.629Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
OPT 1.3B-Knowledge Unlearning for Mitigating Privacy Risks in Language Models
Gemma-2 2B0.721Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
Test-Time Fine-Tuning with SIFT + GPT-2 (124M)0.862Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
Phi-3 7B0.678Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
GPT-Neo 2.7B-Knowledge Unlearning for Mitigating Privacy Risks in Language Models
Hybrid H3 125M-Hungry Hungry Hippos: Towards Language Modeling with State Space Models
Larger Transformer 771M (fine-tuned)-Need a Small Specialized Language Model? Plan Early!-
GPT-Neo 1.3B-Knowledge Unlearning for Mitigating Privacy Risks in Language Models
Larger Transformer 771M (pre-trained)-Need a Small Specialized Language Model? Plan Early!-
Test-Time Fine-Tuning with SIFT + GPT-2 (774M)0.762Efficiently Learning at Test-Time: Active Fine-Tuning of LLMs-
GPT-2 Medium 355M (pre-trained)1.0928The Pile: An 800GB Dataset of Diverse Text for Language Modeling
GPT-2 XL 1.5B (pre-trained)1.0468The Pile: An 800GB Dataset of Diverse Text for Language Modeling
0 of 39 row(s) selected.