trustyai-explainability/trustyai-explainability-python-examples

Examples for the Python bindings for TrustyAI's explainability library

37
/ 100
Emerging

When you're working with complex AI models, especially those used for critical decisions like loan approvals or medical diagnoses, it can be hard to understand why a model made a specific prediction. This project helps you peek inside your AI model to see what factors influenced its output. It takes your model's predictions and shows you the key drivers, making AI more transparent for data scientists, machine learning engineers, and risk managers.

No commits in the last 6 months.

Use this if you need to understand the reasoning behind your AI model's predictions, assess its fairness, or debug unexpected behaviors.

Not ideal if you are looking for a tool to build AI models from scratch, rather than explain existing ones.

AI explainability machine learning auditing model interpretability AI fairness predictive analytics
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

13

Forks

6

Language

License

Apache-2.0

Last pushed

Oct 07, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/trustyai-explainability/trustyai-explainability-python-examples"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.