HyperAI
HyperAI초신경
홈
뉴스
최신 연구 논문
튜토리얼
데이터셋
백과사전
SOTA
LLM 모델
GPU 랭킹
컨퍼런스
전체 검색
소개
한국어
HyperAI
HyperAI초신경
Toggle sidebar
전체 사이트 검색...
⌘
K
홈
SOTA
시각적 연역
Visual Entailment On Snli Ve Val
Visual Entailment On Snli Ve Val
평가 지표
Accuracy
평가 결과
이 벤치마크에서 각 모델의 성능 결과
Columns
모델 이름
Accuracy
Paper Title
Repository
CLIP-ViL
80.20
How Much Can CLIP Benefit Vision-and-Language Tasks?
-
EVE-ROI*
70.81
Visual Entailment: A Novel Task for Fine-Grained Image Understanding
-
OFA
91.0
OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
-
SimVLM
86.21
SimVLM: Simple Visual Language Model Pretraining with Weak Supervision
-
UNITER
78.98
UNITER: UNiversal Image-TExt Representation Learning
-
Prompt Tuning
90.04
Prompt Tuning for Generative Multimodal Pretrained Models
-
VILLA-LARGE
80.18
Large-Scale Adversarial Training for Vision-and-Language Representation Learning
-
CoCa
87.0
CoCa: Contrastive Captioners are Image-Text Foundation Models
-
SOHO
85.00
Seeing Out of tHe bOx: End-to-End Pre-training for Vision-Language Representation Learning
-
0 of 9 row(s) selected.
Previous
Next
Visual Entailment On Snli Ve Val | SOTA | HyperAI초신경