jiangnanboy/llm_security

利用分类法和敏感词检测法对生成式大模型的输入和输出内容进行安全检测,尽早识别风险内容。The input and output contents of generative large model are checked by classification method and sensitive word detection method to identify content risk as early as possible.

35
/ 100
Emerging

This tool helps ensure that content generated by large language models (LLMs) is safe and appropriate before it reaches your users. It takes the text inputs and outputs of your generative AI models and identifies potentially risky content, helping you catch issues early. It's designed for anyone managing or deploying LLM applications who needs to maintain content standards and mitigate reputation risks.

No commits in the last 6 months.

Use this if you are using generative AI models and need an automated way to screen both user prompts and AI-generated responses for sensitive or harmful content.

Not ideal if you are looking for a tool to analyze general text for sentiment or topic modeling, rather than specific safety and risk detection related to LLMs.

AI-content-moderation generative-AI-safety LLM-risk-management digital-ethics content-filtering
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 12 / 25

How are scores calculated?

Stars

28

Forks

4

Language

Java

License

MIT

Last pushed

Sep 09, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/jiangnanboy/llm_security"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.