HyperAI

Natural Language Inference On Rcb

Metrics

Accuracy
Average F1

Results

Performance results of various models on this benchmark

Comparison Table
Model NameAccuracyAverage F1
Model 10.4180.302
Model 20.5180.357
Model 30.5460.406
Model 40.4630.367
Model 50.4980.306
russiansuperglue-a-russian-language0.7020.68
Model 70.5090.333
Model 8 0.4840.417
Model 90.4730.356
Model 100.4470.408
Model 110.4520.371
Model 120.4450.367
mt5-a-massively-multilingual-pre-trained-text0.4540.366
Model 140.50.356
Model 150.4860.351
Model 160.4680.307
unreasonable-effectiveness-of-rule-based0.4380.4
unreasonable-effectiveness-of-rule-based0.3740.319
Model 190.4610.372
Model 200.4840.452
russiansuperglue-a-russian-language0.4410.301
unreasonable-effectiveness-of-rule-based0.4840.217