HyperAI
Accueil
Actualités
Articles de recherche récents
Tutoriels
Ensembles de données
Wiki
SOTA
Modèles LLM
Classement GPU
Événements
Recherche
À propos
Français
HyperAI
Toggle sidebar
Rechercher sur le site...
⌘
K
Accueil
SOTA
Visual Entailment
Visual Entailment On Snli Ve Val
Visual Entailment On Snli Ve Val
Métriques
Accuracy
Résultats
Résultats de performance de divers modèles sur ce benchmark
Columns
Nom du modèle
Accuracy
Paper Title
Repository
CLIP-ViL
80.20
How Much Can CLIP Benefit Vision-and-Language Tasks?
EVE-ROI*
70.81
Visual Entailment: A Novel Task for Fine-Grained Image Understanding
OFA
91.0
OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
SimVLM
86.21
SimVLM: Simple Visual Language Model Pretraining with Weak Supervision
UNITER
78.98
UNITER: UNiversal Image-TExt Representation Learning
Prompt Tuning
90.04
Prompt Tuning for Generative Multimodal Pretrained Models
VILLA-LARGE
80.18
Large-Scale Adversarial Training for Vision-and-Language Representation Learning
CoCa
87.0
CoCa: Contrastive Captioners are Image-Text Foundation Models
SOHO
85.00
Seeing Out of tHe bOx: End-to-End Pre-training for Vision-Language Representation Learning
0 of 9 row(s) selected.
Previous
Next