LLM Firewall Defense LLM Tools
Tools for real-time detection and blocking of prompt injection, jailbreaks, and malicious tool calls in LLM applications. Does NOT include vulnerability research, fuzzing frameworks, or governance/compliance tooling.
There are 73 llm firewall defense tools tracked. 3 score above 50 (established tier). The highest-rated is ethz-spylab/agentdojo at 69/100 with 471 stars. 2 of the top 10 are actively maintained.
Get all 73 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=llm-firewall-defense&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
ethz-spylab/agentdojo
A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents. |
|
Established |
| 2 |
guardrails-ai/guardrails
Adding guardrails to large language models. |
|
Established |
| 3 |
JasonLovesDoggo/caddy-defender
Caddy module to block or manipulate requests originating from AIs or cloud... |
|
Established |
| 4 |
deadbits/vigil-llm
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky... |
|
Emerging |
| 5 |
inkdust2021/VibeGuard
Uses just 1% memory while protecting 99% of your personal privacy. |
|
Emerging |
| 6 |
Heiberg-Industries/designbrief
Design guardrails, not templates. A library of UI design direction files... |
|
Emerging |
| 7 |
dropbox/llm-security
Dropbox LLM Security research code and results |
|
Emerging |
| 8 |
AmenRa/GuardBench
A Python library for guardrail models evaluation. |
|
Emerging |
| 9 |
whitecircle-ai/circle-guard-bench
First-of-its-kind AI benchmark for evaluating the protection capabilities of... |
|
Emerging |
| 10 |
pampanic/pam_panic
A PAM module that protects sensitive data and provides a panic function for... |
|
Emerging |
| 11 |
lukeslp/ux-oss-safeguard
Content safety evaluator built on OpenAI's gpt-oss-safeguard-20b — zero... |
|
Emerging |
| 12 |
Zierax/Basic-ML-prompt-injections
llm attacks basic payloads |
|
Emerging |
| 13 |
turbot/guardrails-lib-ai
A unified interface for multiple AI language model providers, tailored for... |
|
Emerging |
| 14 |
North-Shore-AI/LlmGuard
AI Firewall and guardrails for LLM-based Elixir applications |
|
Emerging |
| 15 |
yihedeng9/DuoGuard
DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails |
|
Emerging |
| 16 |
henchiyb/breaker-ai
Breaker AI - Security check for your LLM prompts |
|
Emerging |
| 17 |
automorphic-ai/aegis
Self-hardening firewall for large language models |
|
Emerging |
| 18 |
kdunee/intentguard
A Python library for verifying code properties using natural language assertions. |
|
Emerging |
| 19 |
k14uz/PhishNet
PhishNet is an experimental research project implementing Reinforced... |
|
Emerging |
| 20 |
tyoung1996/guardrail-layer
Guardrail Layer: Open-source AI data privacy firewall — redact, audit, and... |
|
Emerging |
| 21 |
upss-standard/universal-prompt-security-standard
Universal Prompt Security Standard (UPSS): A framework for externalizing,... |
|
Emerging |
| 22 |
microsoft/llmail-inject-challenge-analysis
Data Analysis of the results of llmail-inject challenge |
|
Experimental |
| 23 |
recurprotocol/recur-protocol
Self-evolving recursive AI sentinels for real-time IP protection and prompt... |
|
Experimental |
| 24 |
willshacklett/gvai-safety-systems
Runtime AI safety & security infrastructure for monitoring constraint strain... |
|
Experimental |
| 25 |
acebot712/promptguard-node
PromptGuard Node.js SDK — Drop-in security for AI applications |
|
Experimental |
| 26 |
SkintLabs/WonderwallAi
WonderwallAi — Open-source AI firewall SDK for LLM applications. Prompt... |
|
Experimental |
| 27 |
chenglin1112/AgentTrust
Real-time trustworthiness evaluation and safety interception for AI agents.... |
|
Experimental |
| 28 |
Agentic-AI-Risk-Mitigation/Janus
System-level security for LLM agents: fine-grained policy enforcement on... |
|
Experimental |
| 29 |
logicbunchhq/ai_guardrails
🛡️ The safety & validation layer for LLMs in Ruby. Prevents JSON errors,... |
|
Experimental |
| 30 |
heymumford/cognilateral-trust
AI that tells you when it's guessing. Confidence tiers + accountability for... |
|
Experimental |
| 31 |
AUTHENSOR/prompt-injection-benchmark
Standardized benchmark for testing AI safety scanners. Run your scanner, get... |
|
Experimental |
| 32 |
studiomeyer-io/ai-shield
LLM security toolkit — prompt injection detection, PII masking, cost... |
|
Experimental |
| 33 |
marcin-jasinski/outlier-vanguard
Prompt injection detection via novelty scoring. Securing agentic AI systems... |
|
Experimental |
| 34 |
isartor-ai/Isartor
Pure-Rust Prompt Firewall that eliminates unnecessary LLM cloud calls.... |
|
Experimental |
| 35 |
personal-vault/pvp
Personal Vault Project |
|
Experimental |
| 36 |
voodooEntity/ghost_trap
Multi‑surface anti‑scraping: GitHub Action appends a README trap;... |
|
Experimental |
| 37 |
danielmaddaleno/llm-guardrails-toolkit
Pluggable guardrails pipeline for LLM apps – PII redaction, prompt... |
|
Experimental |
| 38 |
zjdtm/secure-prompt-spring-boot-starter
Spring Boot LLM Prompt Injection Defense Library |
|
Experimental |
| 39 |
urcuqui/PhishAwareBot
PhishAwareBot is a web application that generates and displays two types of... |
|
Experimental |
| 40 |
sammm0308/bonklm
Enforce security guardrails for large language models in Node.js... |
|
Experimental |
| 41 |
darkmailr/darkmailr
darkmailr - Offline Phishing Simulation Tool |
|
Experimental |
| 42 |
christopherpaquin/Guardrails-AI
Reliable AI Infrastructure: Input/Output validation, structured data... |
|
Experimental |
| 43 |
maro-style/SlangShield
SlangShield: Because ‘hey bro’ doesn’t belong in Q4 stakeholder updates (IT/EN). |
|
Experimental |
| 44 |
ogulcanaydogan/Prompt-Injection-Firewall
Real-time prompt injection detection and prevention middleware for LLM... |
|
Experimental |
| 45 |
MoeinAlvandi/sovereign-vault
🛡️ Automate your data backups with Sovereign Vault, ensuring secure 3-2-1... |
|
Experimental |
| 46 |
michusSq/configguard
Detect network misconfigurations using AI to ensure compliance with security... |
|
Experimental |
| 47 |
Rami8612/signguard-ai
Protect yourself and your team before signing — designed for multi-signature... |
|
Experimental |
| 48 |
zentinelproxy/zentinel-agent-ai-gateway
AI Gateway agent for Zentinel proxy - prompt injection, PII, jailbreak detection |
|
Experimental |
| 49 |
maltyxx/guardix
An autonomous Web Application Firewall (WAF) that uses a Large Language... |
|
Experimental |
| 50 |
DrPwner/PromptSniffer
PromptSniffer is a security auditing tool designed for authorized... |
|
Experimental |
| 51 |
JuanAirala/freedom
🔒 Protect your data with Freedom, an encrypted storage app using AES-256-GCM... |
|
Experimental |
| 52 |
wwa/FIMjector
FIMjector is an exploit for OpenAI GPT models based on Fill-In-the-Middle... |
|
Experimental |
| 53 |
yashcanbuild/social_shield
🛡️ AI-powered email security platform that detects phishing attempts,... |
|
Experimental |
| 54 |
GPierce9/Vault-Protocol-v2.6-Safer-AI-by-Design
Vault Protocol is an integrated, trauma-informed safety architecture built... |
|
Experimental |
| 55 |
cybertechajju/LLM-PROMPT-INJECTION-PAYLOAD-S
Unlock safe, high-signal prompt workflows for ethical hacking and AI red-teaming |
|
Experimental |
| 56 |
SH-Nihil-Mukkesh-25/Vaultify
Vaultify is an intelligent security system combining ESP32 hardware with... |
|
Experimental |
| 57 |
miksto/danger-llm_validator
Danger plugin for write rules in natural language, and let an LLM ensure... |
|
Experimental |
| 58 |
Nibir1/VaultSim
An event-driven AI social engineering simulator. Execute prompt injection... |
|
Experimental |
| 59 |
fitsblb/PhishGuardAI
Modern, explainable phishing URL detection with FastAPI, policy bands, and... |
|
Experimental |
| 60 |
arihantprasad07/guni
AI agent security middleware — detect prompt injection, phishing and goal... |
|
Experimental |
| 61 |
metawake/puppetry-detector
**Puppetry Detector** is a modular engine for detecting structured and... |
|
Experimental |
| 62 |
shaqir/healthcare-ai-safety-demo-
Healthcare AI safety demo — 7-layer architecture with prompt injection... |
|
Experimental |
| 63 |
shaqir/healthcare-ai-safety
Healthcare AI safety system — 7-layer architecture with prompt injection... |
|
Experimental |
| 64 |
augustyatuhsexpeimentation/agentshield
Security firewall for AI agents — block prompt injection, data exfiltration,... |
|
Experimental |
| 65 |
mhsn1/ghostshield
AI-powered LLM security scanner, real prompt injection attacks. |
|
Experimental |
| 66 |
Tearfullnex/SpecGuard
🛡️ Enforce AI behavior guidelines with SpecGuard, a tool that turns policies... |
|
Experimental |
| 67 |
57karakalkan/MetaSafe-Guardian-
🛡️ Monitor virtual spaces with AI to detect harmful behavior, manage safe... |
|
Experimental |
| 68 |
RahulR767/defender-acl-blocker
🔒 Block Microsoft Defender's user space with easy ACL adjustments to enhance... |
|
Experimental |
| 69 |
ilicitos/agent-creds
🔐 Inject API credentials securely into unmodified code via a transparent... |
|
Experimental |
| 70 |
juyterman1000/llm-safety
Stop prompt injections in 20ms. The safety toolkit every LLM app needs. No... |
|
Experimental |
| 71 |
rizkycsv/PromptGuard
🔒 Safeguard LLM behavior with PromptGuard to detect unseen regressions and... |
|
Experimental |
| 72 |
vartulzeroshieldai/AIGuardX
Loss prevention and policy enforcement for generative AI tools (ChatGPT,... |
|
Experimental |
| 73 |
nehamanoj1105/Banking-LLM-Security
Detection and Mitigation of Prompt Injection and Jailbreaking Attacks in... |
|
Experimental |