jiangnanboy/llm_security
利用分类法和敏感词检测法对生成式大模型的输入和输出内容进行安全检测,尽早识别风险内容。The input and output contents of generative large model are checked by classification method and sensitive word detection method to identify content risk as early as possible.
This tool helps ensure that content generated by large language models (LLMs) is safe and appropriate before it reaches your users. It takes the text inputs and outputs of your generative AI models and identifies potentially risky content, helping you catch issues early. It's designed for anyone managing or deploying LLM applications who needs to maintain content standards and mitigate reputation risks.
No commits in the last 6 months.
Use this if you are using generative AI models and need an automated way to screen both user prompts and AI-generated responses for sensitive or harmful content.
Not ideal if you are looking for a tool to analyze general text for sentiment or topic modeling, rather than specific safety and risk detection related to LLMs.
Stars
28
Forks
4
Language
Java
License
MIT
Category
Last pushed
Sep 09, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/jiangnanboy/llm_security"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
OWASP/www-project-top-10-for-large-language-model-applications
OWASP Top 10 for Large Language Model Apps (Part of the GenAI Security Project)
esbmc/esbmc-ai
Automated Code Repair suite powered by ESBMC and LLMs.
cla7aye15I4nd/PatchAgent
[USENIX Security 25] PatchAgent is a LLM-based practical program repair agent that mimics human...
iSEngLab/AwesomeLLM4APR
[TOSEM 2026]A Systematic Literature Review on Large Language Models for Automated Program Repair
YerbaPage/MGDebugger
Multi-Granularity LLM Debugger [ICSE2026]