UKPLab/nessie
Automatically detect errors in annotated corpora.
When you're building datasets by hand for machine learning, this tool helps you automatically find mistakes in your annotated text, tokens, or spans. It takes your existing labeled data and identifies potentially incorrect annotations, allowing you to prioritize which instances to review and correct. This is for anyone who creates or manages human-annotated text datasets, such as data scientists, linguists, or quality assurance specialists.
No commits in the last 6 months.
Use this if you need to improve the quality and efficiency of your human data annotation process by automatically flagging potential errors in your datasets.
Not ideal if you are looking for a tool to perform the initial annotation of your data, as this is designed for error detection in already-annotated corpora.
Stars
48
Forks
6
Language
Python
License
MIT
Category
Last pushed
Sep 08, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/UKPLab/nessie"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
chakki-works/seqeval
A Python framework for sequence labeling evaluation(named-entity recognition, pos tagging, etc...)
Hironsan/anago
Bidirectional LSTM-CRF and ELMo for Named-Entity Recognition, Part-of-Speech Tagging and so on.
jbesomi/texthero
Text preprocessing, representation and visualization from zero to hero.
hamelsmu/ktext
Utilities for preprocessing text for deep learning with Keras
asahi417/tner
Language model fine-tuning on NER with an easy interface and cross-domain evaluation. "T-NER: An...