promptfoo/promptfoo-action

The GitHub Action for Promptfoo. Test your prompts, agents, and RAGs. AI Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.

54
/ 100
Established

This GitHub Action helps AI developers, ML engineers, and prompt engineers automatically test their large language model (LLM) prompts and RAG (Retrieval Augmented Generation) systems. When you modify prompts in your code, it evaluates the changes and provides a "before/after" comparison directly in your pull request. This allows you to quickly see how prompt edits impact model performance and identify regressions or improvements.

Use this if you are developing with LLMs and need to ensure that changes to your prompts or RAG configurations don't negatively impact performance, providing continuous evaluation and feedback.

Not ideal if you are a business user or an individual not directly involved in the coding and deployment of LLM-powered applications.

LLM development Prompt engineering AI red teaming Model evaluation Continuous integration
No Package No Dependents
Maintenance 10 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 20 / 25

How are scores calculated?

Stars

47

Forks

23

Language

TypeScript

License

MIT

Last pushed

Mar 12, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/promptfoo/promptfoo-action"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.