License:
CC-BY-4.0
Steward:
0DIN by MozillaTask: N/A
Release Date: 5/8/2026
Format: JSONL
Size: 22.39 KB
Share
0DIN, the 0Day Investigative Network, was founded by Mozilla in 2024 to reward responsible researchers for their efforts in securing GenAI models. This dataset is a weekly export of the public 0DIN disclosures published at https://0din.ai/disclosures, in accordance with the 0DIN Research Terms and Disclosure Policy. Each record corresponds to a single validated, published disclosure and is grounded in 0DIN's public research frameworks: Security Boundaries (prompt extraction, guardrail jailbreak, interpreter jailbreak, content manipulation, weights/layers disclosure, prompt injection), the Jailbreak Taxonomy (category → strategy → technique), the Social Impact Score (SIS, levels 1–5), and the Nude Imagery Rating System (NIRS, levels 1–5). Records include title, summary, severity, security boundary, taxonomy triplets, affected models and vendors, visible test results, SIS and NIRS scores when assessed, public researcher credit when supplied, reference URLs, and disclosure/publication timestamps. Prompts, model responses, attack payloads, detection signatures, variant prompts, and submitter PII (other than a self-supplied public credit string) are intentionally excluded. Use the data to study GenAI vulnerability trends, train and evaluate safety classifiers, build detection pipelines, or inform model-card transparency on known weaknesses.
Licensing
Creative Commons Attribution 4.0 International (CC-BY-4.0)
https://spdx.org/licenses/CC-BY-4.0.htmlRestrictions/Special Constraints
Use is governed by CC-BY-4.0 (https://creativecommons.org/licenses/by/4.0/): attribute 0DIN (https://0din.ai) when redistributing or building on this dataset. The dataset is intended for security research, AI safety evaluation, education, and the development of defensive tooling, consistent with 0DIN's Research Terms and Disclosure Policy. Do not use it to conduct unauthorized attacks on production systems, and do not attempt to re-identify, contact, or attribute disclosures to individuals beyond the public researcher-credit field.
Forbidden Usage
- Conducting unauthorized attacks on production AI systems or third-party services. - Re-identifying or attempting to deanonymize researchers, including those listed as "Anonymous". - Training models to evade safety guardrails or content moderation in ways that violate the targeted system's terms of service. - Generating content that is illegal in the jurisdictions where the data is processed (e.g., CSAM, NCII). - Harassing any individual, organization, or vendor referenced in the data.
Ethical Review
Each record reflects a disclosure already publicly visible on 0din.ai, vetted by 0DIN staff before publication. PII is excluded at the publisher layer; only researcher credit explicitly supplied by the submitter is included. Anonymous submissions are preserved as "Anonymous". Suppressed disclosures are excluded by 0DIN's publishing pipeline. The dataset is a structured replica of public-facing content, not a release of internal data.
Intended Use
Training and evaluation of AI safety classifiers; automated red-teaming and detection pipelines; trend analysis across GenAI security boundaries and jailbreak taxonomies; model-card transparency on known weaknesses; reproducible academic security research with citable public disclosures.
Bundle layout: a single .tar.gz archive containing three files at its root — vulnerabilities.jsonl (one disclosure per line), manifest.json (generated_at, dataset_name, dataset_schema_version, record_count, source URL, publisher_git_sha), and README.md. Records are ordered by published_at descending, then uuid ascending. The current schema version is recorded in the manifest.