Natural Language Inference On Rcb
Metrics
Accuracy
Average F1
Results
Performance results of various models on this benchmark
Comparison Table
Model Name | Accuracy | Average F1 |
---|---|---|
Model 1 | 0.418 | 0.302 |
Model 2 | 0.518 | 0.357 |
Model 3 | 0.546 | 0.406 |
Model 4 | 0.463 | 0.367 |
Model 5 | 0.498 | 0.306 |
russiansuperglue-a-russian-language | 0.702 | 0.68 |
Model 7 | 0.509 | 0.333 |
Model 8 | 0.484 | 0.417 |
Model 9 | 0.473 | 0.356 |
Model 10 | 0.447 | 0.408 |
Model 11 | 0.452 | 0.371 |
Model 12 | 0.445 | 0.367 |
mt5-a-massively-multilingual-pre-trained-text | 0.454 | 0.366 |
Model 14 | 0.5 | 0.356 |
Model 15 | 0.486 | 0.351 |
Model 16 | 0.468 | 0.307 |
unreasonable-effectiveness-of-rule-based | 0.438 | 0.4 |
unreasonable-effectiveness-of-rule-based | 0.374 | 0.319 |
Model 19 | 0.461 | 0.372 |
Model 20 | 0.484 | 0.452 |
russiansuperglue-a-russian-language | 0.441 | 0.301 |
unreasonable-effectiveness-of-rule-based | 0.484 | 0.217 |