HyperAI

Referring Expression Segmentation On Refcoco 4

Metriken

Overall IoU

Ergebnisse

Leistungsergebnisse verschiedener Modelle zu diesem Benchmark

Modellname
Overall IoU
Paper TitleRepository
C3VG77.96Multi-task Visual Grounding with Coarse-to-Fine Consistency Constraints
UniLSeg-10078.29Universal Segmentation at Arbitrary Granularity with Language Instruction
STEP (5-fold)52.33See-Through-Text Grouping for Referring Image Segmentation-
ReLA71.02GRES: Generalized Referring Expression Segmentation
MaskRIS (Swin-B)74.46MaskRIS: Semantic Distortion-aware Data Augmentation for Referring Image Segmentation
DETRIS78.6Densely Connected Parameter-Efficient Tuning for Referring Image Segmentation
CRIS68.08CRIS: CLIP-Driven Referring Image Segmentation
MagNet71.32Mask Grounding for Referring Image Segmentation
VLT68.43VLT: Vision-Language Transformer and Query Generation for Referring Segmentation
SHNet58.46Comprehensive Multi-Modal Interactions for Referring Image Segmentation
GROUNDHOG75.0GROUNDHOG: Grounding Large Language Models to Holistic Segmentation-
LAVT68.38LAVT: Language-Aware Vision Transformer for Referring Image Segmentation
MaIL65.92MaIL: A Unified Mask-Image-Language Trimodal Network for Referring Image Segmentation-
HyperSeg83.5HyperSeg: Towards Universal Visual Segmentation with Large Language Model
SafaRi-B74.53SafaRi:Adaptive Sequence Transformer for Weakly Supervised Referring Expression Segmentation-
MattNet52.39MAttNet: Modular Attention Network for Referring Expression Comprehension
CPMC53.44Referring Image Segmentation via Cross-Modal Progressive Comprehension
PolyFormer-B72.89PolyFormer: Referring Image Segmentation as Sequential Polygon Generation
VATEX-Vision-Aware Text Features in Referring Image Segmentation: From Object Understanding to Context Understanding
CMSA47.60Cross-Modal Self-Attention Network for Referring Image Segmentation
0 of 29 row(s) selected.