rhesis-ai/rhesis

Open-source platform & SDK for testing LLM and agentic apps. Define expected behavior, generate and run test scenarios, and review failures collaboratively.

58
/ 100
Established

This platform helps teams ensure their AI applications, like chatbots or intelligent agents, work correctly and safely before they go live. You provide plain language requirements and context, and it generates comprehensive test scenarios. The output shows if your AI meets expectations, avoids harmful content, and retains information, making it useful for product managers, domain experts, and engineers building AI-powered products.

296 stars. Available on PyPI.

Use this if you need a collaborative way to test your LLM or agentic applications against defined requirements and potential vulnerabilities.

Not ideal if you are looking for a post-production monitoring solution rather than a pre-production validation tool.

AI-product-development conversational-AI-testing LLM-quality-assurance AI-safety-testing agentic-system-validation
Maintenance 10 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 13 / 25

How are scores calculated?

Stars

296

Forks

21

Language

Python

License

Last pushed

Mar 13, 2026

Commits (30d)

0

Dependencies

1

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/rhesis-ai/rhesis"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.