shapash and xai

These tools are competitors, as both Shapash and EthicalML/xai offer frameworks and toolboxes for generating explanations and interpretations of machine learning models.

shapash
70
Verified
xai
64
Established
Maintenance 13/25
Adoption 11/25
Maturity 25/25
Community 21/25
Maintenance 6/25
Adoption 10/25
Maturity 25/25
Community 23/25
Stars: 3,150
Forks: 373
Downloads:
Commits (30d): 3
Language: Jupyter Notebook
License: Apache-2.0
Stars: 1,229
Forks: 186
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No risk flags
No risk flags

About shapash

MAIF/shapash

🔅 Shapash: User-friendly Explainability and Interpretability to Develop Reliable and Transparent Machine Learning Models

This project helps data scientists and machine learning engineers understand why their predictive models make certain decisions. It takes a trained machine learning model and its input data, then generates easy-to-understand visualizations and reports that explain the model's behavior. The output helps both technical and non-technical stakeholders gain trust and insights into the model's predictions.

machine-learning-auditing model-explanation data-science-communication predictive-analytics AI-transparency

About xai

EthicalML/xai

XAI - An eXplainability toolbox for machine learning

This tool helps data scientists and machine learning engineers analyze and evaluate their machine learning models to ensure fairness and transparency. It takes in your dataset and trained model, then outputs visualizations and metrics that highlight data imbalances, feature importance, and model performance across different groups. This is for anyone building or deploying machine learning models who needs to understand why their model makes certain decisions and identify potential biases.

Machine Learning Ethics Bias Detection Model Auditing Data Fairness AI Governance

Scores updated daily from GitHub, PyPI, and npm data. How scores work