yueliu1999/Awesome-Jailbreak-on-LLMs
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, datasets, evaluations, and analyses.
Organizes attack and defense methods across multiple threat vectors—targeting reasoning models, black-box/white-box scenarios, multi-turn conversations, RAG systems, and multi-modal inputs—alongside guardrail approaches like learning-based defenses and guard models. The repository indexes implementation code and datasets alongside paper citations, enabling reproducible comparison of attack/defense effectiveness. Covers emerging safety challenges in reasoning-heavy LLMs (o1-style models) and multimodal systems alongside traditional text-based jailbreaks.
1,245 stars. Actively maintained with 15 commits in the last 30 days.
Stars
1,245
Forks
101
Language
—
License
MIT
Category
Last pushed
Mar 07, 2026
Commits (30d)
15
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/yueliu1999/Awesome-Jailbreak-on-LLMs"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
wuyoscar/ISC-Bench
Internal Safety Collapse: Turning LLMs into a "Jailbroken State" Without "a Jailbreak Attack".
xirui-li/DrAttack
Official implementation of paper: DrAttack: Prompt Decomposition and Reconstruction Makes...
yiksiu-chan/SpeakEasy
[ICML 2025] Speak Easy: Eliciting Harmful Jailbreaks from LLMs with Simple Interactions
tmlr-group/DeepInception
[arXiv:2311.03191] "DeepInception: Hypnotize Large Language Model to Be Jailbreaker"
Techiral/awesome-llm-jailbreaks
Latest AI Jailbreak Payloads & Exploit Techniques for GPT, QWEN, and all LLM Models