Agenta-AI/agenta
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM observability all in one place.
This platform helps product and engineering teams build reliable applications powered by Large Language Models (LLMs). It provides tools to refine the prompts that guide LLMs, test their performance with various inputs, and monitor how they behave once deployed. You can input different prompts and test cases, then analyze the LLM's responses and performance metrics.
3,923 stars. Actively maintained with 322 commits in the last 30 days.
Use this if you are building LLM-powered applications and need a systematic way to manage prompts, evaluate model responses, and observe performance in production.
Not ideal if you are solely using LLMs for simple, one-off tasks and do not require iterative development, testing, or production monitoring.
Stars
3,923
Forks
492
Language
TypeScript
License
—
Category
Last pushed
Mar 13, 2026
Commits (30d)
322
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/Agenta-AI/agenta"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Recent Releases
Compare
Related tools
langfuse/langfuse
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management,...
Arize-ai/phoenix
AI Observability & Evaluation
Mirascope/mirascope
The LLM Anti-Framework
Helicone/helicone
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
algorithmicsuperintelligence/optillm
Optimizing inference proxy for LLMs