marcelovca90/auto-ml-evaluation

Code of the article "A practical evaluation of AutoML tools for binary, multiclass, and multilabel classification".

31
/ 100
Emerging

When you need to choose the best automated machine learning (AutoML) tool for your classification problem, this project helps by comparing 16 popular tools across different types of classification tasks (binary, multiclass, multilabel) and 21 real-world datasets. It provides a comprehensive evaluation of how each tool performs in terms of accuracy and speed. Data scientists and machine learning engineers can use the results to inform their tool selection for specific projects.

No commits in the last 6 months.

Use this if you are a data scientist or ML engineer struggling to choose the most effective AutoML tool for a classification task and need objective performance benchmarks.

Not ideal if you are looking for an AutoML tool to directly apply to your dataset, rather than an evaluation of existing tools.

machine-learning-evaluation classification autoML-selection data-science-workflow predictive-modeling
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

9

Forks

1

Language

Jupyter Notebook

License

MIT

Last pushed

May 15, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/marcelovca90/auto-ml-evaluation"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.