HyperAI

MORE Multimodal Object-Entity Relation Extraction Dataset

Date

7 months ago

Size

2.86 GB

Organization

Publish URL

github.com

MORE (A Multimodal Object-Entity Relation Extraction Dataset) is a multimodal object-entity relationship extraction dataset proposed by the National Key Laboratory of Nanjing University in 2024. The related paper results are "MORE: A Multimodal Object-Entity Relation Extraction Dataset with a Benchmark Evaluation".

This dataset combines text and image information to provide a complex challenge for machine learning models, namely how to accurately extract entities from text and establish correct relationships with visual objects in images. The MORE dataset contains 21 different relationship types, covering 20,264 multimodal relationship facts, which are annotated on 3,559 pairs of text captions and corresponding images. Each fact in the dataset involves entities identified from text and objects detected from images, which requires the model to not only understand the text content, but also be able to recognize and understand the image content. In addition, the dataset contains 13,520 visual objects, with an average of 3.8 objects per image.

An example of MORE, including 2 entities from text and several visual objects from images, generates 2 new multimodal relation facts.

MORE.torrent
Seeding 1Downloading 0Completed 123Total Downloads 103
  • MORE/
    • README.md
      1.84 KB
    • README.txt
      3.67 KB
      • data/
        • A-Multimodal-Object-Entity-Relation-Extraction-Dataset.zip
          2.86 GB