HyperAI

Common Sense Reasoning On Rwsd

Metrics

Accuracy

Results

Performance results of various models on this benchmark

Comparison Table
Model NameAccuracy
Model 10.669
Model 20.571
Model 30.669
russiansuperglue-a-russian-language0.662
russiansuperglue-a-russian-language0.84
Model 60.636
Model 70.649
Model 80.545
Model 90.669
Model 100.675
mt5-a-massively-multilingual-pre-trained-text0.669
Model 120.669
unreasonable-effectiveness-of-rule-based0.669
Model 140.669
unreasonable-effectiveness-of-rule-based0.597
Model 160.669
Model 170.662
Model 180.669
Model 190.669
unreasonable-effectiveness-of-rule-based0.669
Model 210.669
Model 220.669