lucagioacchini/auto-pen-bench
This repo contains the codes of the penetration test benchmark for Generative Agents presented in the paper "AutoPenBench: Benchmarking Generative Agents for Penetration Testing". It contains also the instructions to install, develop and test new vulnerable containers to include in the benchmark.
This benchmark helps cybersecurity researchers and professionals evaluate how well AI-powered penetration testing agents can find vulnerabilities in simulated systems. It takes a generative AI agent and a definition of a vulnerable machine, then reports on the agent's ability to identify and exploit weaknesses. Security analysts, red teamers, and AI researchers focused on offensive security would use this to assess agent performance.
Use this if you are developing or evaluating generative AI agents for automated penetration testing and need a standardized way to measure their effectiveness against various vulnerabilities.
Not ideal if you are looking for a tool to perform actual penetration tests on live production systems or if you are not working with generative AI agents.
Stars
68
Forks
20
Language
Python
License
MIT
Category
Last pushed
Oct 28, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/lucagioacchini/auto-pen-bench"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
GoogleCloudPlatform/vertex-ai-samples
Notebooks, code samples, sample apps, and other resources that demonstrate how to use, develop...
neo4j-partners/hands-on-lab-neo4j-and-google
Hands on Lab for Neo4j and Google
lynnlangit/learning-cloud
Courses, sample code, articles & screencasts - AWS, Azure, & GCP
GoogleCloudPlatform/applied-ai-engineering-samples
This repository compiles code samples and notebooks demonstrating how to use Generative AI on...
streamlit/30DaysOfAI
30 Days of AI