HyperAI

Natural Language Inference On Snli

Metriken

% Test Accuracy
% Train Accuracy
Parameters

Ergebnisse

Leistungsergebnisse verschiedener Modelle zu diesem Benchmark

Modellname
% Test Accuracy
% Train Accuracy
Parameters
Paper TitleRepository
RoBERTa-large + self-explaining layer92.3?355m+Self-Explaining Structures Improve NLP Models
Distance-based Self-Attention Network86.389.64.7mDistance-based Self-Attention Network for Natural Language Inference-
Stacked Bi-LSTMs (shortcut connections, max-pooling, attention)84.4--Combining Similarity Features and Deep Representation Learning for Stance Detection in the Context of Checking Fake News
300D Gumbel TreeLSTM encoders85.691.22.9mLearning to Compose Task-Specific Tree Structures
MT-DNN91.697.2330mMulti-Task Deep Neural Networks for Natural Language Understanding
CBS-1 + ESIM86.73--Parameter Re-Initialization through Cyclical Batch Size Schedules-
SJRC (BERT-Large +SRL)91.395.7308mExplicit Contextual Semantics for Text Comprehension-
CA-MTL92.192.6340mConditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
1024D GRU encoders w/ unsupervised 'skip-thoughts' pre-training81.498.815mOrder-Embeddings of Images and Language
200D decomposable attention model with intra-sentence attention86.890.5580kA Decomposable Attention Model for Natural Language Inference
MT-DNN-SMART_0.1%ofTrainingData---SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
600D (300+300) BiLSTM encoders with intra-attention and symbolic preproc.85.085.92.8mLearning Natural Language Inference using Bidirectional LSTM model and Inner-Attention
600D BiLSTM with generalized pooling86.694.965mEnhancing Sentence Embedding with Generalized Pooling
Enhanced Sequential Inference Model (Chen et al., [2017a])88.0--Enhanced LSTM for Natural Language Inference
300D Reinforced Self-Attention Network86.392.63.1mReinforced Self-Attention Network: a Hybrid of Hard and Soft Attention for Sequence Modeling
100D DF-LSTM84.685.2320k--
600D (300+300) Deep Gated Attn. BiLSTM encoders85.590.512mRecurrent Neural Network-Based Sentence Encoder with Gated Attention for Natural Language Inference
450D DR-BiLSTM88.594.17.5mDR-BiLSTM: Dependent Reading Bidirectional LSTM for Natural Language Inference-
300D Residual stacked encoders85.789.89.7mShortcut-Stacked Sentence Encoders for Multi-Domain Inference
ESIM + ELMo Ensemble89.392.140mDeep contextualized word representations
0 of 98 row(s) selected.