AhsanAyub/malicious-prompt-detection

Detection of malicious prompts used to exploit large language models (LLMs) by leveraging supervised machine learning classifiers.

29
/ 100
Experimental

This project helps developers and engineers building applications powered by large language models (LLMs) to identify and block malicious prompts. It takes user input prompts and classifies them as either 'benign' or 'malicious' to prevent prompt injection attacks. This is for engineers and developers responsible for the security and robustness of their LLM-based applications.

No commits in the last 6 months.

Use this if you are building an application that uses large language models and need to protect it from prompt injection attacks.

Not ideal if you are a non-technical user looking for a ready-to-use content moderation tool for general text inputs, rather than LLM-specific security.

LLM security prompt engineering AI application development cybersecurity machine learning engineering
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 8 / 25
Community 15 / 25

How are scores calculated?

Stars

20

Forks

4

Language

Python

License

Category

ai-red-teaming

Last pushed

Oct 30, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/AhsanAyub/malicious-prompt-detection"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.