Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
-
Updated
Jun 10, 2025 - Python
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
[NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.
First-of-its-kind AI benchmark for evaluating the protection capabilities of large language model (LLM) guard systems (guardrails and safeguards)
Implementation of paper 'Defending Large Language Models against Jailbreak Attacks via Semantic Smoothing'
Add a description, image, and links to the llm-jailbreaks topic page so that developers can more easily learn about it.
To associate your repository with the llm-jailbreaks topic, visit your repo's landing page and select "manage topics."