wuyoscar/ISC-Bench
Internal Safety Collapse: Turning LLMs into a "Jailbroken State" Without "a Jailbreak Attack".
Provides both single-turn templates and autonomous agent-based execution modes for systematically triggering safety failures in frontier LLMs through incomplete professional workflows (e.g., biomedical analysis, chemical synthesis) rather than adversarial prompts. The benchmark uses domain-specific task templates across biology, chemistry, and epidemiology with layered evaluation scaffolds, paired with an ISC Arena leaderboard tracking vulnerability patterns across models via API-based reproducibility and community submissions through GitHub Issues.
677 stars. Actively maintained with 289 commits in the last 30 days.
Stars
677
Forks
127
Language
Python
License
—
Category
Last pushed
Mar 28, 2026
Commits (30d)
289
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/wuyoscar/ISC-Bench"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
yueliu1999/Awesome-Jailbreak-on-LLMs
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods...
yiksiu-chan/SpeakEasy
[ICML 2025] Speak Easy: Eliciting Harmful Jailbreaks from LLMs with Simple Interactions
xirui-li/DrAttack
Official implementation of paper: DrAttack: Prompt Decomposition and Reconstruction Makes...
tmlr-group/DeepInception
[arXiv:2311.03191] "DeepInception: Hypnotize Large Language Model to Be Jailbreaker"
Techiral/awesome-llm-jailbreaks
Latest AI Jailbreak Payloads & Exploit Techniques for GPT, QWEN, and all LLM Models