Audio Captioning On Clotho
Metrics
BLEU-4
CIDEr
METEOR
ROUGE-L
Results
Performance results of various models on this benchmark
Comparison Table
Model Name | BLEU-4 | CIDEr | METEOR | ROUGE-L |
---|---|---|---|---|
valor-vision-audio-language-omni-perception | 16.2 | 0.423 | 17.4 | 38.2 |
audio-captioning-using-gated-recurrent-units | - | 0.18 | - | - |
vast-a-vision-audio-subtitle-text-omni-1 | 19 | 0.519 | 19.3 | 40.8 |
the-sjtu-system-for-dcase2021-challenge-task | - | 0.468 | - | - |
the-dcase-2021-challenge-task-6-system | - | 0.400 | - | - |
the-ntt-dcase2020-challenge-task-6-system | - | 0.319 | - | - |
qwen-audio-advancing-universal-audio | - | 0.441 | - | - |
slam-aac-enhancing-audio-captioning-with | - | 0.515 | 0.197 | - |
enhancing-automated-audio-captioning-via | - | 0.513 | 0.197 | - |
audio-flamingo-a-novel-audio-language-model | 17.4 | 0.489 | 18.7 | 39.4 |