HKU-TASR/Imperio
[IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.
This project helps security researchers and AI auditors understand a new type of vulnerability in machine learning models, specifically in image classification. It takes a clean image dataset and, using language-guided instructions, trains a 'backdoored' model. The output is a model that can be controlled to misclassify specific images based on text commands, while still performing accurately on normal inputs.
No commits in the last 6 months.
Use this if you are researching advanced backdoor attacks on image classification models and need a tool to create and evaluate language-guided backdoor vulnerabilities.
Not ideal if you are looking for a defensive tool to detect or mitigate existing backdoors, or if your focus is on NLP model vulnerabilities.
Stars
44
Forks
4
Language
Python
License
MIT
Category
Last pushed
Feb 18, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/HKU-TASR/Imperio"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
GreyDGL/PentestGPT
Automated Penetration Testing Agentic Framework Powered by Large Language Models
berylliumsec/nebula
AI-powered penetration testing assistant for automating recon, note-taking, and vulnerability analysis.
ipa-lab/hackingBuddyGPT
Helping Ethical Hackers use LLMs in 50 Lines of Code or less..
MorDavid/BruteForceAI
Advanced LLM-powered brute-force tool combining AI intelligence with automated login attacks
mbrg/power-pwn
An offensive/defense security toolset for discovery, recon and ethical assessment of AI Agents