HyperAI
HyperAI초신경
홈
플랫폼
문서
뉴스
연구 논문
튜토리얼
데이터셋
백과사전
SOTA
LLM 모델
GPU 랭킹
컨퍼런스
전체 검색
소개
서비스 약관
개인정보 처리방침
한국어
HyperAI
HyperAI초신경
Toggle Sidebar
전체 사이트 검색...
⌘
K
Command Palette
Search for a command to run...
플랫폼
홈
SOTA
의미적 텍스트 유사성
Semantic Textual Similarity On Mrpc
Semantic Textual Similarity On Mrpc
평가 지표
F1
평가 결과
이 벤치마크에서 각 모델의 성능 결과
Columns
모델 이름
F1
Paper Title
BigBird
91.5
Big Bird: Transformers for Longer Sequences
T5-3B
92.5
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
MobileBERT
-
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
BERT-Base
-
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning
Charformer-Tall
91.4
Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
RoBERTa-large 355M + Entailment as Few-shot Learner
91.0
Entailment as Few-Shot Learner
Nyströmformer
88.1%
Nyströmformer: A Nyström-Based Algorithm for Approximating Self-Attention
SMART-BERT
-
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
RoBERTa-large 355M (MLP quantized vector-wise, fine-tuned)
-
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
FNet-Large
-
FNet: Mixing Tokens with Fourier Transforms
SqueezeBERT
-
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?
XLNet (single model)
-
XLNet: Generalized Autoregressive Pretraining for Language Understanding
SMART
-
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
T5-Large
92.4
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
TinyBERT-6 67M
-
TinyBERT: Distilling BERT for Natural Language Understanding
TinyBERT-4 14.5M
-
TinyBERT: Distilling BERT for Natural Language Understanding
DistilBERT 66M
-
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
ERNIE 2.0 Base
-
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding
T5-Small
89.7
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Q8BERT (Zafrir et al., 2019)
-
Q8BERT: Quantized 8Bit BERT
0 of 45 row(s) selected.
Previous
Next
Semantic Textual Similarity On Mrpc | SOTA | HyperAI초신경