HyperAIHyperAI

Command Palette

Search for a command to run...

Llama-Nemotron-Post-Training-Dataset Post-training Dataset

Date

2 months ago

Organization

NVIDIA

Paper URL

2505.00949

License

CC BY 4.0

Llama-Nemotron-Post-Training-Dataset is a large-scale post-training dataset open sourced by NVIDIA in 2025. The related paper results are "Llama-Nemotron: Efficient Reasoning Models", which aims to improve the mathematics, code, general reasoning and instruction following capabilities of the Llama-Nemotron series models in the post-training stage (such as SFT and RL).

This dataset combines data from supervised fine-tuning (SFT) and reinforcement learning (RL) phases. The current version, v1.1 (which adds approximately 2.2 million math samples and 500,000 code reasoning samples compared to the previous version), is suitable for training AI agents, chatbots, RAG systems, and other AI-driven applications.

Data distribution (by number of category entries)

  • Mathematics: 22,066,397
  • Code: 10,108,883
  • Science: 708,920
  • Instruction following: 56,339
  • Chat: 39,792
  • Safety: 31,426

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp