agencyenterprise/PromptInject
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
PromptInject helps you understand how robust your large language model (LLM) is against malicious user inputs that try to hijack its intended purpose or leak its internal instructions. It takes your LLM and various attack prompts as input, then shows you how easily the model can be misled. This tool is for AI safety researchers, machine learning engineers, and product managers responsible for deploying LLMs in real-world applications.
465 stars. No commits in the last 6 months. Available on PyPI.
Use this if you are deploying a customer-facing LLM and need to rigorously test its security and reliability against adversarial inputs.
Not ideal if you are looking for a general-purpose tool to improve LLM performance or fine-tune models for specific tasks.
Stars
465
Forks
44
Language
Python
License
MIT
Category
Last pushed
Feb 26, 2024
Commits (30d)
0
Dependencies
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/agencyenterprise/PromptInject"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
protectai/llm-guard
The Security Toolkit for LLM Interactions
MaxMLang/pytector
Easy to use LLM Prompt Injection Detection / Detector Python Package with support for local...
utkusen/promptmap
a security scanner for custom LLM applications
Resk-Security/Resk-LLM
Resk is a robust Python library designed to enhance security and manage context when...
Dicklesworthstone/acip
The Advanced Cognitive Inoculation Prompt