ezhilezhil/Toxic-Command-Detection
Toxic command detection refers to the process of identifying harmful or inappropriate language in text, particularly in the context of interactions with generative AI and large language models.
No commits in the last 6 months.
Stars
—
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Nov 27, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/ezhilezhil/Toxic-Command-Detection"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
zake7749/DeepToxic
top 1% solution to toxic comment classification challenge on Kaggle.
DenisIndenbom/AntiToxicBot
AntiToxicBot is a bot that detects toxics in a chat using Data Science and Machine Learning...
aralroca/react-text-toxicity
Detect text toxicity in a simple way, using React. Based in a Keras model, loaded with Tensorflow.js.
bensonruan/Toxic-Comment-Classifier
Toxic-Comment-Classifier
charliegerard/safe-space
Github action that checks the toxicity level of comments and PR reviews to help make repos safe spaces.