gcunhase/NLPMetrics
Python code for various NLP metrics
This tool helps machine translation researchers and practitioners evaluate the quality of their automated translations. You input a machine-generated translation and one or more human-generated reference translations, and it calculates metrics like BLEU, GLEU, and WER. The output is a score indicating how well the machine translation matches the human references, helping you assess and improve your models.
169 stars. No commits in the last 6 months.
Use this if you need to quantitatively measure the performance of machine translation systems against human-quality translations.
Not ideal if you need metrics for tasks beyond machine translation, such as text summarization or image captioning, which are not yet fully implemented.
Stars
169
Forks
28
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Oct 17, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/gcunhase/NLPMetrics"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
google/langfun
OO for LLMs
tanaos/artifex
Small Language Model Inference, Fine-Tuning and Observability. No GPU, no labeled data needed.
preligens-lab/textnoisr
Adding random noise to a text dataset, and controlling very accurately the quality of the result
vulnerability-lookup/VulnTrain
A tool to generate datasets and models based on vulnerabilities descriptions from @Vulnerability-Lookup.
masakhane-io/masakhane-mt
Machine Translation for Africa