yihedeng9/DuoGuard

DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails

34
/ 100
Emerging

This project helps developers and engineers ensure the safety and appropriateness of responses generated by large language models (LLMs) in multiple languages. It takes an LLM's output in various languages and determines if it contains unsafe or undesirable content. This is useful for AI developers, product managers, or content moderation teams building or deploying multilingual LLMs.

No commits in the last 6 months.

Use this if you need to automatically detect and flag unsafe content from your LLMs across multiple languages with high accuracy and efficiency.

Not ideal if you are looking for a tool to generate text or translate content, as its primary function is safety moderation.

AI-safety content-moderation LLM-deployment multilingual-AI responsible-AI
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

32

Forks

4

Language

Python

License

Apache-2.0

Last pushed

Feb 26, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/yihedeng9/DuoGuard"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.