HyperAIHyperAI

Command Palette

Search for a command to run...

UltraSafety Large Model Safety Evaluation Dataset

Date

2 years ago

Size

17.43 MB

Organization

Tsinghua University

The UltraSafety dataset was jointly created by Renmin University, Tsinghua University, and Tencent to evaluate and improve the security of large models. UltraSafety exports 1,000 safe seed instructions from AdvBench and MaliciousInstruct, and uses Self-Instruct to guide another 2,000 instructions. The research team manually screened the jailbreak prompts in AutoDAN and finally screened out 830 high-quality jailbreak prompts. UltraSafety contains a total of 3,000 harmful instructions, each with a related jailbreak prompt. Each harmful instruction corresponds to the completion result generated by our model at different security levels, and is accompanied by a rating specified by GPT4, where a rating of 1 means harmless and a rating of 0 means harmful. The UltraSafety dataset aims to assist researchers in training models that can identify and prevent potential security threats through these detailed security-related instructions.

UltraSafety.torrent
Seeding 2Downloading 1Completed 374Total Downloads 839
  • UltraSafety/
    • README.md
      1.53 KB
    • README.txt
      3.07 KB
      • data/
        • UltraSafety.zip
          17.43 MB

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp