GitsSaikat/Guardian-Agent

Improving AI Systems with Self-Defense Mechanisms

39
/ 100
Emerging

This project helps AI system developers protect their AI models from malicious prompts that try to hijack or manipulate their behavior. It takes in user prompts and ensures the AI model only acts on legitimate, safe instructions, preventing it from being 'jailbroken' or manipulated. Developers building AI agents who want to maintain strict control over their AI's responses and actions would use this.

No commits in the last 6 months.

Use this if you are developing an AI agent and need to build in robust self-defense mechanisms against adversarial prompt attacks and jailbreaking attempts.

Not ideal if you are an end-user of an existing AI system, as this tool is for developers to integrate into their AI agent's architecture.

AI safety AI security prompt engineering AI agent development malicious prompt prevention
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 17 / 25

How are scores calculated?

Stars

23

Forks

8

Language

Python

License

MIT

Last pushed

Feb 28, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/GitsSaikat/Guardian-Agent"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.