HyperAI

Semantic Textual Similarity On Mrpc

Metriken

F1

Ergebnisse

Leistungsergebnisse verschiedener Modelle zu diesem Benchmark

Modellname
F1
Paper TitleRepository
BigBird91.5Big Bird: Transformers for Longer Sequences
T5-3B92.5Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
MobileBERT-MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
BERT-Base-Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning-
Charformer-Tall91.4Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
RoBERTa-large 355M + Entailment as Few-shot Learner91.0Entailment as Few-Shot Learner
Nyströmformer88.1%Nyströmformer: A Nyström-Based Algorithm for Approximating Self-Attention
SMART-BERT-SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
RoBERTa-large 355M (MLP quantized vector-wise, fine-tuned)-LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
FNet-Large-FNet: Mixing Tokens with Fourier Transforms
SqueezeBERT-SqueezeBERT: What can computer vision teach NLP about efficient neural networks?
XLNet (single model)-XLNet: Generalized Autoregressive Pretraining for Language Understanding
SMART-SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
T5-Large92.4Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
TinyBERT-6 67M-TinyBERT: Distilling BERT for Natural Language Understanding
TinyBERT-4 14.5M-TinyBERT: Distilling BERT for Natural Language Understanding
DistilBERT 66M-DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
ERNIE 2.0 Base-ERNIE 2.0: A Continual Pre-training Framework for Language Understanding
T5-Small89.7Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Q8BERT (Zafrir et al., 2019)-Q8BERT: Quantized 8Bit BERT
0 of 45 row(s) selected.