HyperAI초신경
홈
뉴스
최신 연구 논문
튜토리얼
데이터셋
백과사전
SOTA
LLM 모델
GPU 랭킹
컨퍼런스
전체 검색
소개
한국어
HyperAI초신경
Toggle sidebar
전체 사이트 검색...
⌘
K
홈
SOTA
Zero Shot Transfer Image Classification
Zero Shot Transfer Image Classification On 1
Zero Shot Transfer Image Classification On 1
평가 지표
Accuracy (Private)
평가 결과
이 벤치마크에서 각 모델의 성능 결과
Columns
모델 이름
Accuracy (Private)
Paper Title
Repository
EVA-CLIP-E/14+
82
EVA-CLIP: Improved Training Techniques for CLIP at Scale
CWCL
-
-
-
EVA-CLIP-18B
83.8
EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters
-
CLIP(ViT-L/14-336px)
76.2
Learning Transferable Visual Models From Natural Language Supervision
LiT ViT-e
85.4
PaLI: A Jointly-Scaled Multilingual Language-Image Model
CLIP
-
Learning Transferable Visual Models From Natural Language Supervision
REACT
78.5
Learning Customized Visual Models with Retrieval-Augmented Knowledge
PaLI
72.11
PaLI: A Jointly-Scaled Multilingual Language-Image Model
InternVL-C
83.2
InternVL: Scaling up Vision Foundation Models and Aligning for Generic Visual-Linguistic Tasks
CLIP (ResNet50)
59.6
Learning Transferable Visual Models From Natural Language Supervision
CoCa
86.3
CoCa: Contrastive Captioners are Image-Text Foundation Models
BASIC (Lion)
88.3
-
-
M2-Encoder
88.5
M2-Encoder: Advancing Bilingual Image-Text Understanding by Large-scale Efficient Pretraining
IMP-MoE-L
83.9
Alternating Gradient Descent and Mixture-of-Experts for Integrated Multimodal Perception
-
LiT-22B
85.9
Scaling Vision Transformers to 22 Billion Parameters
CLIPA (ViT-H/14-336px)
81.8
-
-
ALIGN
76.4
Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
MAWS (ViT-2B)
82.1
The effectiveness of MAE pre-pretraining for billion-scale pretraining
AltCLIP
74.5
AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities
Diffusion Classifier (zero-shot)
61.4
Your Diffusion Model is Secretly a Zero-Shot Classifier
0 of 23 row(s) selected.
Previous
Next