HyperAI

Common Sense Reasoning On Rwsd

Metriken

Accuracy

Ergebnisse

Leistungsergebnisse verschiedener Modelle zu diesem Benchmark

Vergleichstabelle
ModellnameAccuracy
Modell 10.669
Modell 20.571
Modell 30.669
russiansuperglue-a-russian-language0.662
russiansuperglue-a-russian-language0.84
Modell 60.636
Modell 70.649
Modell 80.545
Modell 90.669
Modell 100.675
mt5-a-massively-multilingual-pre-trained-text0.669
Modell 120.669
unreasonable-effectiveness-of-rule-based0.669
Modell 140.669
unreasonable-effectiveness-of-rule-based0.597
Modell 160.669
Modell 170.662
Modell 180.669
Modell 190.669
unreasonable-effectiveness-of-rule-based0.669
Modell 210.669
Modell 220.669