Aaryanverma/trustifai

TrustifAI: A Comprehensive Framework for AI Trustworthiness

42
/ 100
Emerging

This helps evaluate the reliability of AI-generated responses from Large Language Models (LLMs) and RAG systems. It takes your AI's answer, the original question, and any supporting documents, then provides a detailed 'Trust Score' and visual explanations. This is for AI product managers, content quality assurance specialists, or anyone who needs to ensure their AI's outputs are accurate, consistent, and well-supported.

Available on PyPI.

Use this if you need to systematically assess and demonstrate the trustworthiness of your AI's outputs beyond simple correctness, especially for critical applications.

Not ideal if you only need a basic 'yes/no' answer on whether your AI is factually correct, or if you're not using LLMs or RAG systems.

AI-quality-assurance LLM-evaluation RAG-system-auditing content-accuracy AI-trust-metrics
Maintenance 10 / 25
Adoption 5 / 25
Maturity 20 / 25
Community 7 / 25

How are scores calculated?

Stars

10

Forks

1

Language

Python

License

MIT

Last pushed

Feb 08, 2026

Commits (30d)

0

Dependencies

10

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/Aaryanverma/trustifai"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.