rhesis-ai/rhesis
Open-source platform & SDK for testing LLM and agentic apps. Define expected behavior, generate and run test scenarios, and review failures collaboratively.
This platform helps teams ensure their AI applications, like chatbots or intelligent agents, work correctly and safely before they go live. You provide plain language requirements and context, and it generates comprehensive test scenarios. The output shows if your AI meets expectations, avoids harmful content, and retains information, making it useful for product managers, domain experts, and engineers building AI-powered products.
296 stars. Available on PyPI.
Use this if you need a collaborative way to test your LLM or agentic applications against defined requirements and potential vulnerabilities.
Not ideal if you are looking for a post-production monitoring solution rather than a pre-production validation tool.
Stars
296
Forks
21
Language
Python
License
—
Category
Last pushed
Mar 13, 2026
Commits (30d)
0
Dependencies
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/rhesis-ai/rhesis"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Related tools
openvinotoolkit/model_server
A scalable inference server for models optimized with OpenVINO™
madroidmaq/mlx-omni-server
MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically...
NVIDIA-NeMo/Guardrails
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based...
generative-computing/mellea
Mellea is a library for writing generative programs.
taco-group/OpenEMMA
OpenEMMA, a permissively licensed open source "reproduction" of Waymo’s EMMA model.