Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
-
Updated
Feb 3, 2026 - Python
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
[NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.
A working POC of a GPT-5 jailbreak via PROMISQROUTE (Prompt-based Router Open-Mode Manipulation) with a barebones C2 server & agent generation demo.
First-of-its-kind AI benchmark for evaluating the protection capabilities of large language model (LLM) guard systems (guardrails and safeguards)
LMAP (large language model mapper) is like NMAP for LLM, is an LLM Vulnerability Scanner and Zero-day Vulnerability Fuzzer.
Implementation of paper 'Defending Large Language Models against Jailbreak Attacks via Semantic Smoothing'
[ICML 2025] Speak Easy: Eliciting Harmful Jailbreaks from LLMs with Simple Interactions
Jailbreak Evaluation Framework -- 2025 Graduate Design for HFUT
🔍 Benchmark jailbreak resilience in LLMs with JailBench for clear insights and improved model defenses against jailbreak attempts.
Chain-of-thought hijacking via template token injection for LLM censorship bypass (GPT-OSS)
Benchmark LLM jailbreak resilience across providers with standardized tests, adversarial mode, rich analytics, and a clean Web UI.
Systematic LLM jailbreak taxonomy — 40 attack patterns, 10 categories, empirical evaluation across 4 frontier models. AI safety research with responsible disclosure.
LLM Jailbreaking via Prompt Rewriting
RetardBench is an open, no-censorship benchmark that ranks large language models purely on how retarded they are.
PESU I/O The Hacker's Gauntlet 24-hours CTF
The Self-Hosted AI Firewall & Gateway. Drop-in guardrails for LLMs running entirely on CPU. Blocks jailbreaks, enforces policies, and ensures compliance in real-time
Debugged version for Tree of Attacks: Jailbreaking Black-Box LLMs Automatically paper and added GPU optimization.
Add a description, image, and links to the llm-jailbreaks topic page so that developers can more easily learn about it.
To associate your repository with the llm-jailbreaks topic, visit your repo's landing page and select "manage topics."