HyperAIHyperAI

Command Palette

Search for a command to run...

Natural Language Inference On Terra

Metrics

Accuracy

Results

Performance results of various models on this benchmark

Paper Title
Human Benchmark0.92RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark
Golden Transformer0.871-
ruRoberta-large finetune0.801-
ruT5-large-finetune0.747-
ruBert-large finetune0.704-
ruBert-base finetune0.703-
ruT5-base-finetune0.692-
RuGPT3Large0.654-
RuBERT plain0.642-
RuBERT conversational0.64-
SBERT_Large_mt_ru_finetuning0.637-
SBERT_Large0.637-
Multilingual Bert0.617-
YaLM 1.0B few-shot0.605-
RuGPT3XL few-shot0.573-
MT5 Large0.561mT5: A massively multilingual pre-trained text-to-text transformer
heuristic majority0.549Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks
majority_class0.513Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks
RuGPT3Medium0.505-
RuGPT3Small0.488-
0 of 22 row(s) selected.