cgnorthcutt/rankpruning

🧹 Formerly for binary classification with noisy labels. Replaced by cleanlab.

41
/ 100
Emerging

This project offers a method for improving the accuracy of binary classification models when your training data has errors in the labels. It takes your existing data (features) and potentially mislabeled training outcomes, and outputs a more reliable model for predicting future outcomes. A research scientist or machine learning practitioner working with noisy datasets would find this useful.

No commits in the last 6 months.

Use this if you are a research scientist trying to reproduce the results of the 2017 UAI publication, 'Learning with Confident Examples: Rank Pruning for Binary Classification with Noisy Labels'.

Not ideal if you are looking for a general-purpose solution for noisy labels; the `cleanlab` package is a more robust and updated alternative for broader applications.

machine-learning-research binary-classification noisy-data model-training academic-research
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

83

Forks

14

Language

Jupyter Notebook

License

MIT

Last pushed

May 15, 2022

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/cgnorthcutt/rankpruning"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.