HyperAIHyperAI

Command Palette

Search for a command to run...

MM-RLHF Multimodal Preference Alignment Dataset

Date

9 months ago

Size

55.33 GB

Organization

Paper URL

arxiv.org

License

Apache 2.0

MM-RLHF (Multimodal Reinforcement Learning from Human Feedback) is a high-quality, fine-grained multimodal dataset.MM-RLHF: The Next Step Forward in Multimodal LLM Alignment", first published on arXiv in 2025 by the Institute of Automation, Chinese Academy of Sciences (CASIA). This dataset aims to promote the alignment research of multimodal large language models (MLLMs) and solve the problems of truthfulness, safety, and alignment with human preferences in practical applications.

The dataset contains 120,000 pairs of fine-grained, manually annotated preference comparison data, covering three areas: image understanding, video analysis, and multimodal security. The amount of data far exceeds existing resources, covering more than 100,000 multimodal task instances. Each piece of data has been carefully scored and interpreted by more than 50 annotators to ensure the high quality and granularity of the data.

Dataset Example

MM-RLHF.torrent
Seeding 2Downloading 0Completed 147Total Downloads 168
  • MM-RLHF/
    • README.md
      1.55 KB
    • README.txt
      3.09 KB
      • data/
        • MM-RLHF.zip
          55.33 GB

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
MM-RLHF Multimodal Preference Alignment Dataset | Datasets | HyperAI