HyperAI
HyperAI
Home
News
Latest Papers
Tutorials
Datasets
Wiki
SOTA
LLM Models
GPU Leaderboard
Events
Search
About
English
HyperAI
HyperAI
Toggle sidebar
Search the site…
⌘
K
Home
SOTA
Unsupervised Semantic Segmentation with Language-image Pre-training
Unsupervised Semantic Segmentation With 9
Unsupervised Semantic Segmentation With 9
Metrics
mIoU
Results
Performance results of various models on this benchmark
Columns
Model Name
mIoU
Paper Title
Repository
MaskCLIP
16.4
Extract Free Dense Labels from CLIP
-
TCL
22.4
Learning to Generate Text-grounded Mask for Open-world Semantic Segmentation from Only Image-Text Pairs
-
TTD (MaskCLIP)
19.4
TTD: Text-Tag Self-Distillation Enhancing Image-Text Alignment in CLIP to Alleviate Single Tag Bias
-
COSMOS ViT-B/16
23.2
COSMOS: Cross-Modality Self-Distillation for Vision Language Pre-training
-
TagAlign
25.3
TagAlign: Improving Vision-Language Alignment with Multi-Tag Classification
-
ReCo
14.8
ReCo: Retrieve and Co-segment for Zero-shot Transfer
-
Trident
28.6
Harnessing Vision Foundation Models for High-Performance, Training-Free Open Vocabulary Segmentation
-
TTD (TCL)
23.7
TTD: Text-Tag Self-Distillation Enhancing Image-Text Alignment in CLIP to Alleviate Single Tag Bias
-
ProxyCLIP
26.8
ProxyCLIP: Proxy Attention Improves CLIP for Open-Vocabulary Segmentation
-
GroupViT
11.1
GroupViT: Semantic Segmentation Emerges from Text Supervision
-
0 of 10 row(s) selected.
Previous
Next
Unsupervised Semantic Segmentation With 9 | SOTA | HyperAI