HyperAIHyperAI

Language Modelling On Wiki 40B

Métriques

Perplexity

Résultats

Résultats de performance de divers modèles sur ce benchmark

Nom du modèle
Perplexity
Paper TitleRepository
FLASH-Quad-8k14.998Transformer Quality in Linear Time-
Combiner-Axial-8k16.49Combiner: Full Attention Transformer with Sparse Computation Cost-
Combiner-Fixed-8k16.60Combiner: Full Attention Transformer with Sparse Computation Cost-
0 of 3 row(s) selected.
Language Modelling On Wiki 40B | SOTA | HyperAI