LLM Firewall Defense LLM Tools

Tools for real-time detection and blocking of prompt injection, jailbreaks, and malicious tool calls in LLM applications. Does NOT include vulnerability research, fuzzing frameworks, or governance/compliance tooling.

There are 73 llm firewall defense tools tracked. 3 score above 50 (established tier). The highest-rated is ethz-spylab/agentdojo at 69/100 with 471 stars. 2 of the top 10 are actively maintained.

Get all 73 projects as JSON

curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=llm-firewall-defense&limit=20"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.

# Tool Score Tier
1 ethz-spylab/agentdojo

A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.

69
Established
2 guardrails-ai/guardrails

Adding guardrails to large language models.

65
Established
3 JasonLovesDoggo/caddy-defender

Caddy module to block or manipulate requests originating from AIs or cloud...

50
Established
4 deadbits/vigil-llm

⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky...

44
Emerging
5 inkdust2021/VibeGuard

Uses just 1% memory while protecting 99% of your personal privacy.

44
Emerging
6 Heiberg-Industries/designbrief

Design guardrails, not templates. A library of UI design direction files...

43
Emerging
7 dropbox/llm-security

Dropbox LLM Security research code and results

42
Emerging
8 AmenRa/GuardBench

A Python library for guardrail models evaluation.

42
Emerging
9 whitecircle-ai/circle-guard-bench

First-of-its-kind AI benchmark for evaluating the protection capabilities of...

40
Emerging
10 pampanic/pam_panic

A PAM module that protects sensitive data and provides a panic function for...

40
Emerging
11 lukeslp/ux-oss-safeguard

Content safety evaluator built on OpenAI's gpt-oss-safeguard-20b — zero...

37
Emerging
12 Zierax/Basic-ML-prompt-injections

llm attacks basic payloads

35
Emerging
13 turbot/guardrails-lib-ai

A unified interface for multiple AI language model providers, tailored for...

35
Emerging
14 North-Shore-AI/LlmGuard

AI Firewall and guardrails for LLM-based Elixir applications

34
Emerging
15 yihedeng9/DuoGuard

DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails

34
Emerging
16 henchiyb/breaker-ai

Breaker AI - Security check for your LLM prompts

34
Emerging
17 automorphic-ai/aegis

Self-hardening firewall for large language models

33
Emerging
18 kdunee/intentguard

A Python library for verifying code properties using natural language assertions.

32
Emerging
19 k14uz/PhishNet

PhishNet is an experimental research project implementing Reinforced...

31
Emerging
20 tyoung1996/guardrail-layer

Guardrail Layer: Open-source AI data privacy firewall — redact, audit, and...

31
Emerging
21 upss-standard/universal-prompt-security-standard

Universal Prompt Security Standard (UPSS): A framework for externalizing,...

30
Emerging
22 microsoft/llmail-inject-challenge-analysis

Data Analysis of the results of llmail-inject challenge

29
Experimental
23 recurprotocol/recur-protocol

Self-evolving recursive AI sentinels for real-time IP protection and prompt...

27
Experimental
24 willshacklett/gvai-safety-systems

Runtime AI safety & security infrastructure for monitoring constraint strain...

26
Experimental
25 acebot712/promptguard-node

PromptGuard Node.js SDK — Drop-in security for AI applications

25
Experimental
26 SkintLabs/WonderwallAi

WonderwallAi — Open-source AI firewall SDK for LLM applications. Prompt...

25
Experimental
27 chenglin1112/AgentTrust

Real-time trustworthiness evaluation and safety interception for AI agents....

25
Experimental
28 Agentic-AI-Risk-Mitigation/Janus

System-level security for LLM agents: fine-grained policy enforcement on...

23
Experimental
29 logicbunchhq/ai_guardrails

🛡️ The safety & validation layer for LLMs in Ruby. Prevents JSON errors,...

23
Experimental
30 heymumford/cognilateral-trust

AI that tells you when it's guessing. Confidence tiers + accountability for...

22
Experimental
31 AUTHENSOR/prompt-injection-benchmark

Standardized benchmark for testing AI safety scanners. Run your scanner, get...

22
Experimental
32 studiomeyer-io/ai-shield

LLM security toolkit — prompt injection detection, PII masking, cost...

22
Experimental
33 marcin-jasinski/outlier-vanguard

Prompt injection detection via novelty scoring. Securing agentic AI systems...

22
Experimental
34 isartor-ai/Isartor

Pure-Rust Prompt Firewall that eliminates unnecessary LLM cloud calls....

22
Experimental
35 personal-vault/pvp

Personal Vault Project

22
Experimental
36 voodooEntity/ghost_trap

Multi‑surface anti‑scraping: GitHub Action appends a README trap;...

22
Experimental
37 danielmaddaleno/llm-guardrails-toolkit

Pluggable guardrails pipeline for LLM apps – PII redaction, prompt...

22
Experimental
38 zjdtm/secure-prompt-spring-boot-starter

Spring Boot LLM Prompt Injection Defense Library

22
Experimental
39 urcuqui/PhishAwareBot

PhishAwareBot is a web application that generates and displays two types of...

22
Experimental
40 sammm0308/bonklm

Enforce security guardrails for large language models in Node.js...

21
Experimental
41 darkmailr/darkmailr

darkmailr - Offline Phishing Simulation Tool

21
Experimental
42 christopherpaquin/Guardrails-AI

Reliable AI Infrastructure: Input/Output validation, structured data...

21
Experimental
43 maro-style/SlangShield

SlangShield: Because ‘hey bro’ doesn’t belong in Q4 stakeholder updates (IT/EN).

21
Experimental
44 ogulcanaydogan/Prompt-Injection-Firewall

Real-time prompt injection detection and prevention middleware for LLM...

21
Experimental
45 MoeinAlvandi/sovereign-vault

🛡️ Automate your data backups with Sovereign Vault, ensuring secure 3-2-1...

21
Experimental
46 michusSq/configguard

Detect network misconfigurations using AI to ensure compliance with security...

21
Experimental
47 Rami8612/signguard-ai

Protect yourself and your team before signing — designed for multi-signature...

21
Experimental
48 zentinelproxy/zentinel-agent-ai-gateway

AI Gateway agent for Zentinel proxy - prompt injection, PII, jailbreak detection

21
Experimental
49 maltyxx/guardix

An autonomous Web Application Firewall (WAF) that uses a Large Language...

21
Experimental
50 DrPwner/PromptSniffer

PromptSniffer is a security auditing tool designed for authorized...

20
Experimental
51 JuanAirala/freedom

🔒 Protect your data with Freedom, an encrypted storage app using AES-256-GCM...

19
Experimental
52 wwa/FIMjector

FIMjector is an exploit for OpenAI GPT models based on Fill-In-the-Middle...

19
Experimental
53 yashcanbuild/social_shield

🛡️ AI-powered email security platform that detects phishing attempts,...

18
Experimental
54 GPierce9/Vault-Protocol-v2.6-Safer-AI-by-Design

Vault Protocol is an integrated, trauma-informed safety architecture built...

17
Experimental
55 cybertechajju/LLM-PROMPT-INJECTION-PAYLOAD-S

Unlock safe, high-signal prompt workflows for ethical hacking and AI red-teaming

17
Experimental
56 SH-Nihil-Mukkesh-25/Vaultify

Vaultify is an intelligent security system combining ESP32 hardware with...

17
Experimental
57 miksto/danger-llm_validator

Danger plugin for write rules in natural language, and let an LLM ensure...

17
Experimental
58 Nibir1/VaultSim

An event-driven AI social engineering simulator. Execute prompt injection...

16
Experimental
59 fitsblb/PhishGuardAI

Modern, explainable phishing URL detection with FastAPI, policy bands, and...

15
Experimental
60 arihantprasad07/guni

AI agent security middleware — detect prompt injection, phishing and goal...

15
Experimental
61 metawake/puppetry-detector

**Puppetry Detector** is a modular engine for detecting structured and...

15
Experimental
62 shaqir/healthcare-ai-safety-demo-

Healthcare AI safety demo — 7-layer architecture with prompt injection...

14
Experimental
63 shaqir/healthcare-ai-safety

Healthcare AI safety system — 7-layer architecture with prompt injection...

14
Experimental
64 augustyatuhsexpeimentation/agentshield

Security firewall for AI agents — block prompt injection, data exfiltration,...

14
Experimental
65 mhsn1/ghostshield

AI-powered LLM security scanner, real prompt injection attacks.

14
Experimental
66 Tearfullnex/SpecGuard

🛡️ Enforce AI behavior guidelines with SpecGuard, a tool that turns policies...

14
Experimental
67 57karakalkan/MetaSafe-Guardian-

🛡️ Monitor virtual spaces with AI to detect harmful behavior, manage safe...

14
Experimental
68 RahulR767/defender-acl-blocker

🔒 Block Microsoft Defender's user space with easy ACL adjustments to enhance...

13
Experimental
69 ilicitos/agent-creds

🔐 Inject API credentials securely into unmodified code via a transparent...

13
Experimental
70 juyterman1000/llm-safety

Stop prompt injections in 20ms. The safety toolkit every LLM app needs. No...

13
Experimental
71 rizkycsv/PromptGuard

🔒 Safeguard LLM behavior with PromptGuard to detect unseen regressions and...

13
Experimental
72 vartulzeroshieldai/AIGuardX

Loss prevention and policy enforcement for generative AI tools (ChatGPT,...

13
Experimental
73 nehamanoj1105/Banking-LLM-Security

Detection and Mitigation of Prompt Injection and Jailbreaking Attacks in...

11
Experimental

Comparisons in this category