ZihanWangKi/CrossWeigh

CrossWeigh: Training Named Entity Tagger from Imperfect Annotations

41
/ 100
Emerging

This project helps researchers and data scientists improve the accuracy of Named Entity Recognition (NER) models when their training data contains human annotation errors. It takes existing NER training datasets that may have mistakes and processes them to identify and down-weight those errors. The output is a more robust NER model that performs better on real-world text analysis tasks, ultimately delivering more accurate entity extraction.

176 stars. No commits in the last 6 months.

Use this if you are training Named Entity Recognition (NER) models and suspect that human errors in your annotated training data are hurting your model's performance.

Not ideal if your primary concern is with evaluating model performance on an already clean and verified test dataset.

natural-language-processing named-entity-recognition text-annotation data-quality machine-learning-training
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 15 / 25

How are scores calculated?

Stars

176

Forks

21

Language

Python

License

Apache-2.0

Last pushed

Jul 25, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/ZihanWangKi/CrossWeigh"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.