HyperAI

Common Sense Reasoning On Rwsd

Métriques

Accuracy

Résultats

Résultats de performance de divers modèles sur ce benchmark

Nom du modèle
Accuracy
Paper TitleRepository
RuBERT conversational0.669--
ruRoberta-large finetune0.571--
ruT5-large-finetune0.669--
Baseline TF-IDF1.10.662RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark
Human Benchmark0.84RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark
RuGPT3Large0.636--
RuGPT3XL few-shot0.649--
Golden Transformer0.545--
Multilingual Bert0.669--
SBERT_Large_mt_ru_finetuning0.675--
MT5 Large0.669mT5: A massively multilingual pre-trained text-to-text transformer
RuGPT3Medium0.669--
heuristic majority0.669Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks-
YaLM 1.0B few-shot0.669--
Random weighted0.597Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks-
RuGPT3Small0.669--
SBERT_Large0.662--
ruBert-large finetune0.669--
ruBert-base finetune0.669--
majority_class0.669Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks-
0 of 22 row(s) selected.