geval-labs/geval
Eval-driven release gates for AI applications
This tool helps AI application teams automate the decision of whether to release a new AI model or prompt. You feed it various performance metrics and business rules, and it consistently produces a clear outcome: 'PASS', 'REQUIRE_APPROVAL', or 'BLOCK'. It's for anyone managing the release of AI features, like AI product managers, machine learning engineers, or release managers.
Use this if you need a consistent, auditable, and automated way to decide on the release readiness of AI application changes based on multiple criteria and signals.
Not ideal if you need an AI system to make the release decisions for you, as this tool only applies your predefined rules.
Stars
14
Forks
6
Language
TypeScript
License
MIT
Category
Last pushed
Feb 23, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/agents/geval-labs/geval"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
StonyBrookNLP/appworld
🌍 AppWorld: A Controllable World of Apps and People for Benchmarking Function Calling and...
qualifire-dev/rogue
AI Agent Evaluator & Red Team Platform
microsoft/WindowsAgentArena
Windows Agent Arena (WAA) 🪟 is a scalable OS platform for testing and benchmarking of...
future-agi/ai-evaluation
Evaluation Framework for all your AI related Workflows
RouteWorks/RouterArena
RouterArena: An open framework for evaluating LLM routers with standardized datasets, metrics,...