chaoluond/safetyllama
Finetune LLaMA-2-7b-chat to perform safety evaluation of user-bot conversation
This project helps ensure AI chatbots provide safe and appropriate responses to user prompts. It takes a conversation between a human and a chatbot as input and outputs an evaluation stating whether the chatbot's response adheres to a set of safety guidelines. AI product managers, trust & safety teams, or developers building AI applications would use this tool.
No commits in the last 6 months.
Use this if you need to automatically monitor and detect unsafe or inappropriate content generated by your AI chatbot before it reaches users.
Not ideal if you are looking for a general-purpose content moderation tool that flags user-generated content directly, rather than chatbot outputs.
Stars
11
Forks
—
Language
Python
License
MIT
Category
Last pushed
Jun 02, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/chaoluond/safetyllama"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Pyenb/Ollama-models
A collection of zipped Ollama models for offline use. Simply download, extract, and set up your...
jaypatel15406/Ollama-Adaptive-Image-Code-Gen
Ollama Adaptive Image Code Gen is an asynchronous Python application that uses LLMs to...
Marvin-VW/python-ollama-local
This Python script enables hands-free interaction with a local Llama2 language model. It...
cognisoc/mullama
Drop-in Ollama replacement. All-in-one local LLM toolkit.
rajkundalia/error-analyzer-with-baml
Analyze Java compilation and runtime errors using BAML with a local Ollama model. This project...