HyperAI

Natural Language Inference On Multinli

Metrics

Matched
Mismatched

Results

Performance results of various models on this benchmark

Model Name
Matched
Mismatched
Paper TitleRepository
ERNIE 2.0 Large88.788.8ERNIE 2.0: A Continual Pre-training Framework for Language Understanding
T5-Base87.186.2Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
ELC-BERT-base 98M (zero init)84.484.5Not all layers are equally as important: Every Layer Counts BERT-
Snorkel MeTaL (ensemble)87.687.2Training Complex Models with Multi-Task Weak Supervision
T5-3B91.491.2Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
GenSen71.471.3Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning
RoBERTa90.8-RoBERTa: A Robustly Optimized BERT Pretraining Approach
Charformer-Tall83.784.4Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
GPST(unsupervised generative syntactic LM)81.882.0Generative Pretrained Structured Transformers: Unsupervised Syntactic Language Models at Scale
LM-CPPF RoBERTa-base--LM-CPPF: Paraphrasing-Guided Data Augmentation for Contrastive Prompt-Based Few-Shot Fine-Tuning
T5-11B-91.7Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
SMART+BERT-BASE--SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
T5-XXL 11B (fine-tuned)92.0-Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
RealFormer86.2886.34RealFormer: Transformer Likes Residual Attention
TinyBERT-6 67M84.683.2TinyBERT: Distilling BERT for Natural Language Understanding
Multi-task BiLSTM + Attn72.272.1GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Bi-LSTM sentence encoder (max-pooling)70.771.1Combining Similarity Features and Deep Representation Learning for Stance Detection in the Context of Checking Fake News
Adv-RoBERTa ensemble91.190.7StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding-
GPT-2-XL 1.5B36.537LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions
UnitedSynT5 (335M)89.8-First Train to Generate, then Generate to Train: UnitedSynT5 for Few-Shot NLI-
0 of 67 row(s) selected.