robinvanschaik/interpret-flair

A small repository to test Captum Explainable AI with a trained Flair transformers-based text classifier.

30
/ 100
Emerging

This project helps machine learning engineers or data scientists understand why a Flair text classification model makes a certain prediction. You provide a trained Flair text classifier and a text input, and it outputs highlighted text showing which words or phrases contributed most to the model's classification decision. It’s for anyone building or evaluating text classification models with Flair who needs to explain their model's reasoning.

No commits in the last 6 months.

Use this if you need to debug, validate, or gain trust in your Flair-based text classification models by understanding their internal workings.

Not ideal if you are looking for a complete, production-ready solution for explainable AI with Flair, as this is an experimental integration.

natural-language-processing machine-learning-engineering model-interpretability text-classification data-science
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 7 / 25

How are scores calculated?

Stars

26

Forks

2

Language

Jupyter Notebook

License

MIT

Last pushed

May 13, 2021

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/robinvanschaik/interpret-flair"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.