ToxiGen
ToxiGen evaluates the propensity of language models to generate toxic content targeting 13 minority groups. It uses adversarially designed prompts to test whether models produce harmful implicit or explicit toxicity.
Metrics
Toxicity rate (%, lower is better)
Created By
Microsoft Research
Paper
View paper →Website
Visit website →Top Model Scores
| Rank | Model | Score | Date |
|---|---|---|---|
| 1 | Claude Opus 4.6 | 1.2% | 2026-02 |
| 2 | GPT-5.2 | 1.8% | 2026-03 |
| 3 | Gemini 3 Ultra | 2.3% | 2026-01 |
| 4 | Llama 4 405B | 3.1% | 2026-01 |
| 5 | Grok 4 | 3.7% | 2026-02 |
Related Safety Benchmarks
TruthfulQA
TruthfulQA measures whether language models generate truthful answers to questions. It includes 817 questions spanning 38 categories where humans might give false answers due to misconceptions, superstitions, or conspiracy theories.
Top: Claude Opus 4.6 — 82.4%
SafetyBench
SafetyBench evaluates the safety of large language models across 7 categories: offensiveness, unfairness and bias, physical health, mental health, illegal activities, ethics and morality, and privacy. It includes questions in both English and Chinese.
Top: Claude Opus 4.6 — 91.7%
TruthfulQA
TruthfulQA measures whether language models generate truthful answers to questions where humans commonly hold misconceptions. The benchmark covers 817 questions across 38 categories including health, law, finance, and conspiracy theories. It specifically targets questions where models are incentivized to reproduce popular falsehoods rather than provide accurate but less common truths, making it a key safety benchmark.
Top: Claude Opus 4.6 — 82.4%