Visual Instruction Following On Llava Bench
Metrics
avg score
Results
Performance results of various models on this benchmark
Comparison Table
Model Name | avg score |
---|---|
sharegpt4v-improving-large-multi-modal-models | 79.9 |
cumo-scaling-multimodal-llm-with-co-upcycled | 85.7 |
improved-baselines-with-visual-instruction | 70.7 |
instructblip-towards-general-purpose-vision | 58.2 |
blip-2-bootstrapping-language-image-pre | 38.1 |
sharegpt4v-improving-large-multi-modal-models | 72.6 |
instructblip-towards-general-purpose-vision | 60.9 |
improved-baselines-with-visual-instruction | 63.4 |