LLM Firewall Defense LLM Tools
Tools for real-time detection and blocking of prompt injection, jailbreaks, and malicious tool calls in LLM applications. Does NOT include vulnerability research, fuzzing frameworks, or governance/compliance tooling.
There are 92 llm firewall defense tools tracked. 2 score above 70 (verified tier). The highest-rated is ethz-spylab/agentdojo at 75/100 with 471 stars and 15,222 monthly downloads. 2 of the top 10 are actively maintained.
Get all 92 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=llm-firewall-defense&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
ethz-spylab/agentdojo
A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents. |
|
Verified |
| 2 |
guardrails-ai/guardrails
Adding guardrails to large language models. |
|
Verified |
| 3 |
JasonLovesDoggo/caddy-defender
Caddy module to block or manipulate requests originating from AIs or cloud... |
|
Emerging |
| 4 |
inkdust2021/VibeGuard
Uses just 1% memory while protecting 99% of your personal privacy. |
|
Emerging |
| 5 |
Heiberg-Industries/designbrief
Design guardrails, not templates. A library of UI design direction files... |
|
Emerging |
| 6 |
ankitlade12/AgentArmor
The full-stack safety layer for AI agents. Budget limits, prompt injection... |
|
Emerging |
| 7 |
AmenRa/GuardBench
A Python library for guardrail models evaluation. |
|
Emerging |
| 8 |
mguard-ai/mguard
Memory defense for AI agents — stops MINJA, AgentPoison, and MemoryGraft... |
|
Emerging |
| 9 |
deadbits/vigil-llm
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky... |
|
Emerging |
| 10 |
whitecircle-ai/circle-guard-bench
First-of-its-kind AI benchmark for evaluating the protection capabilities of... |
|
Emerging |
| 11 |
hexitlabs/vigil
🛡️ Open-source safety guardrail for AI agent tool calls. <2ms, zero dependencies. |
|
Emerging |
| 12 |
turbot/guardrails-lib-ai
A unified interface for multiple AI language model providers, tailored for... |
|
Emerging |
| 13 |
dropbox/llm-security
Dropbox LLM Security research code and results |
|
Emerging |
| 14 |
lukeslp/ux-oss-safeguard
Content safety evaluator built on OpenAI's gpt-oss-safeguard-20b — zero... |
|
Emerging |
| 15 |
mthamil107/prompt-shield
Self-learning prompt injection detection engine that gets smarter with every... |
|
Emerging |
| 16 |
pampanic/pam_panic
A PAM module that protects sensitive data and provides a panic function for... |
|
Emerging |
| 17 |
tyoung1996/guardrail-layer
Guardrail Layer: Open-source AI data privacy firewall — redact, audit, and... |
|
Emerging |
| 18 |
kdunee/intentguard
A Python library for verifying code properties using natural language assertions. |
|
Experimental |
| 19 |
henchiyb/breaker-ai
Breaker AI - Security check for your LLM prompts |
|
Experimental |
| 20 |
North-Shore-AI/LlmGuard
AI Firewall and guardrails for LLM-based Elixir applications |
|
Experimental |
| 21 |
Zierax/Basic-ML-prompt-injections
llm attacks basic payloads |
|
Experimental |
| 22 |
yihedeng9/DuoGuard
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails |
|
Experimental |
| 23 |
upss-standard/universal-prompt-security-standard
Universal Prompt Security Standard (UPSS): A framework for externalizing,... |
|
Experimental |
| 24 |
automorphic-ai/aegis
Self-hardening firewall for large language models |
|
Experimental |
| 25 |
recurprotocol/recur-protocol
Self-evolving recursive AI sentinels for real-time IP protection and prompt... |
|
Experimental |
| 26 |
sleeepeer/PIArena
PIArena: A Platform for Prompt Injection Evaluation |
|
Experimental |
| 27 |
chenglin1112/AgentTrust
Real-time trustworthiness evaluation and safety interception for AI agents.... |
|
Experimental |
| 28 |
k14uz/PhishNet
PhishNet is an experimental research project implementing Reinforced... |
|
Experimental |
| 29 |
Agentic-AI-Risk-Mitigation/Janus
System-level security for LLM agents: fine-grained policy enforcement on... |
|
Experimental |
| 30 |
EctoSpace/EctoLedger
The dashcam and emergency brake for AI agents. A security proxy that... |
|
Experimental |
| 31 |
willshacklett/gvai-safety-systems
Runtime AI safety & security infrastructure for monitoring constraint strain... |
|
Experimental |
| 32 |
microsoft/llmail-inject-challenge-analysis
Data Analysis of the results of llmail-inject challenge |
|
Experimental |
| 33 |
dfghdrtawD/TrustLayer--Security-Control-Plane-For-LLM-AI
🛡️ Secure AI agents with TrustLayer's LLM Firewall, block prompt injection,... |
|
Experimental |
| 34 |
Buddafest/wonderwallai
WonderwallAi — Open-source AI firewall SDK for LLM applications. Prompt... |
|
Experimental |
| 35 |
safe-ai-factory/saifctl
Safety harness for autonomous AI agents: Spec-driven AI factory. Use with... |
|
Experimental |
| 36 |
acebot712/promptguard-node
PromptGuard Node.js SDK — Drop-in security for AI applications |
|
Experimental |
| 37 |
danielmaddaleno/llm-guardrails-toolkit
Pluggable guardrails pipeline for LLM apps – PII redaction, prompt... |
|
Experimental |
| 38 |
zjdtm/secure-prompt-spring-boot-starter
Spring Boot LLM Prompt Injection Defense Library |
|
Experimental |
| 39 |
ogulcanaydogan/Prompt-Injection-Firewall
Real-time prompt injection detection and prevention middleware for LLM... |
|
Experimental |
| 40 |
MoeinAlvandi/sovereign-vault
🛡️ Automate your data backups with Sovereign Vault, ensuring secure 3-2-1... |
|
Experimental |
| 41 |
heymumford/cognilateral-trust
AI that tells you when it's guessing. Confidence tiers + accountability for... |
|
Experimental |
| 42 |
AUTHENSOR/prompt-injection-benchmark
Standardized benchmark for testing AI safety scanners. Run your scanner, get... |
|
Experimental |
| 43 |
studiomeyer-io/ai-shield
LLM security toolkit — prompt injection detection, PII masking, cost... |
|
Experimental |
| 44 |
marcin-jasinski/outlier-vanguard
Prompt injection detection via novelty scoring. Securing agentic AI systems... |
|
Experimental |
| 45 |
mrSamDev/llm-moat
TypeScript toolkit for prompt injection detection, sanitization, and LLM... |
|
Experimental |
| 46 |
isartor-ai/Isartor
Pure-Rust Prompt Firewall that eliminates unnecessary LLM cloud calls.... |
|
Experimental |
| 47 |
michusSq/configguard
Detect network misconfigurations using AI to ensure compliance with security... |
|
Experimental |
| 48 |
WardLink/TrustLayer--Security-Control-Plane-For-LLM-AI
TrustLayer is an API-first security control plane for LLM apps and AI... |
|
Experimental |
| 49 |
AdirD/prompt-security-node
🚀 Unofficial Node.js SDK for Prompt Security's Protection API. |
|
Experimental |
| 50 |
sammm0308/bonklm
Enforce security guardrails for large language models in Node.js... |
|
Experimental |
| 51 |
GaBySuGy/agent-guardrails
🛡️ Enforce AI rules with Agent Guardrails, preventing breaches and ensuring... |
|
Experimental |
| 52 |
logicbunchhq/ai_guardrails
🛡️ The safety & validation layer for LLMs in Ruby. Prevents JSON errors,... |
|
Experimental |
| 53 |
maro-style/SlangShield
SlangShield: Because ‘hey bro’ doesn’t belong in Q4 stakeholder updates (IT/EN). |
|
Experimental |
| 54 |
Rami8612/signguard-ai
Protect yourself and your team before signing — designed for multi-signature... |
|
Experimental |
| 55 |
zentinelproxy/zentinel-agent-ai-gateway
AI Gateway agent for Zentinel proxy - prompt injection, PII, jailbreak detection |
|
Experimental |
| 56 |
christopherpaquin/Guardrails-AI
Reliable AI Infrastructure: Input/Output validation, structured data... |
|
Experimental |
| 57 |
voodooEntity/ghost_trap
Multi‑surface anti‑scraping: GitHub Action appends a README trap;... |
|
Experimental |
| 58 |
new-world-coder/AgentShield
Developer tool for detecting and testing loopholes in agentic AI applications |
|
Experimental |
| 59 |
maltyxx/guardix
An autonomous Web Application Firewall (WAF) that uses a Large Language... |
|
Experimental |
| 60 |
DrPwner/PromptSniffer
PromptSniffer is a security auditing tool designed for authorized... |
|
Experimental |
| 61 |
urcuqui/PhishAwareBot
PhishAwareBot is a web application that generates and displays two types of... |
|
Experimental |
| 62 |
Nibir1/VaultSim
An event-driven AI social engineering simulator. Execute prompt injection... |
|
Experimental |
| 63 |
personal-vault/pvp
Personal Vault Project |
|
Experimental |
| 64 |
GPierce9/Vault-Protocol-v2.6-Safer-AI-by-Design
Vault Protocol is an integrated, trauma-informed safety architecture built... |
|
Experimental |
| 65 |
SH-Nihil-Mukkesh-25/Vaultify
Vaultify is an intelligent security system combining ESP32 hardware with... |
|
Experimental |
| 66 |
57karakalkan/MetaSafe-Guardian-
🛡️ Monitor virtual spaces with AI to detect harmful behavior, manage safe... |
|
Experimental |
| 67 |
darkmailr/darkmailr
darkmailr - Offline Phishing Simulation Tool |
|
Experimental |
| 68 |
dakshaladia/lost-in-the-middle-prompt-injection
Research study on context-window analysis of LLMs |
|
Experimental |
| 69 |
arihantprasad07/guni
AI agent security middleware — detect prompt injection, phishing and goal... |
|
Experimental |
| 70 |
JuanAirala/freedom
🔒 Protect your data with Freedom, an encrypted storage app using AES-256-GCM... |
|
Experimental |
| 71 |
ajutamangdev/PromptShield
PromptShield is an open-source LLM firewall intended to inspect prompts for... |
|
Experimental |
| 72 |
Zikodenomics/cipher-vault
🔐 Manage your passwords securely with Cipher Vault, a simple Python CLI that... |
|
Experimental |
| 73 |
pavani-n-hash/guardianmesh
AI agent security gateway that applies config‑driven rules and LLM‑based... |
|
Experimental |
| 74 |
Tearfullnex/SpecGuard
🛡️ Enforce AI behavior guidelines with SpecGuard, a tool that turns policies... |
|
Experimental |
| 75 |
RahulR767/defender-acl-blocker
🔒 Block Microsoft Defender's user space with easy ACL adjustments to enhance... |
|
Experimental |
| 76 |
rizkycsv/PromptGuard
🔒 Safeguard LLM behavior with PromptGuard to detect unseen regressions and... |
|
Experimental |
| 77 |
vartulzeroshieldai/AIGuardX
Loss prevention and policy enforcement for generative AI tools (ChatGPT,... |
|
Experimental |
| 78 |
shaqir/healthcare-ai-safety-demo-
Healthcare AI safety demo — 7-layer architecture with prompt injection... |
|
Experimental |
| 79 |
shaqir/healthcare-ai-safety
Healthcare AI safety system — 7-layer architecture with prompt injection... |
|
Experimental |
| 80 |
augustyatuhsexpeimentation/agentshield
Security firewall for AI agents — block prompt injection, data exfiltration,... |
|
Experimental |
| 81 |
satrijan/LLM-PROMPT-INJECTION-PAYLOAD-S
🛡️ Explore and test prompt injection techniques safely for AI applications,... |
|
Experimental |
| 82 |
mhsn1/ghostshield
AI-powered LLM security scanner, real prompt injection attacks. |
|
Experimental |
| 83 |
montanaflynn/AdversarialBench
Adversarial prompt-injection benchmark for LLMs |
|
Experimental |
| 84 |
kourgeorge/prompt-sentinel
Python library designed to protect sensitive data when interacting with... |
|
Experimental |
| 85 |
cybertechajju/LLM-PROMPT-INJECTION-PAYLOAD-S
Unlock safe, high-signal prompt workflows for ethical hacking and AI red-teaming |
|
Experimental |
| 86 |
metawake/puppetry-detector
**Puppetry Detector** is a modular engine for detecting structured and... |
|
Experimental |
| 87 |
yashcanbuild/social_shield
🛡️ AI-powered email security platform that detects phishing attempts,... |
|
Experimental |
| 88 |
wwa/FIMjector
FIMjector is an exploit for OpenAI GPT models based on Fill-In-the-Middle... |
|
Experimental |
| 89 |
rohilrg/CatchPromptInjection
This repo focus on how to deal with prompt injection problem faced by LLMs |
|
Experimental |
| 90 |
ilicitos/agent-creds
🔐 Inject API credentials securely into unmodified code via a transparent... |
|
Experimental |
| 91 |
juyterman1000/llm-safety
Stop prompt injections in 20ms. The safety toolkit every LLM app needs. No... |
|
Experimental |
| 92 |
miksto/danger-llm_validator
Danger plugin for write rules in natural language, and let an LLM ensure... |
|
Experimental |