promptfoo/promptfoo-action
The GitHub Action for Promptfoo. Test your prompts, agents, and RAGs. AI Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.
This GitHub Action helps AI developers, ML engineers, and prompt engineers automatically test their large language model (LLM) prompts and RAG (Retrieval Augmented Generation) systems. When you modify prompts in your code, it evaluates the changes and provides a "before/after" comparison directly in your pull request. This allows you to quickly see how prompt edits impact model performance and identify regressions or improvements.
Use this if you are developing with LLMs and need to ensure that changes to your prompts or RAG configurations don't negatively impact performance, providing continuous evaluation and feedback.
Not ideal if you are a business user or an individual not directly involved in the coding and deployment of LLM-powered applications.
Stars
47
Forks
23
Language
TypeScript
License
MIT
Category
Last pushed
Mar 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/promptfoo/promptfoo-action"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related tools
promptfoo/promptfoo
Test your prompts, agents, and RAGs. Red teaming/pentesting/vulnerability scanning for AI....
bigscience-workshop/promptsource
Toolkit for creating, sharing and using natural language prompts.
dair-ai/Prompt-Engineering-Guide
🐙 Guides, papers, lessons, notebooks and resources for prompt engineering, context engineering,...
thunlp/OpenPrompt
An Open-Source Framework for Prompt-Learning.
LearnPrompt/LearnPrompt
永久免费开源的 AIGC 课程, 目前已支持Prompt Engineering, ChatGPT, Midjourney, Runway, Stable Diffusion,...