kavgan/ROUGE-2.0
ROUGE automatic summarization evaluation toolkit. Support for ROUGE-[N, L, S, SU], stemming and stopwords in different languages, unicode text evaluation, CSV output.
This tool helps researchers and developers evaluate how good their automatically generated text summaries or translations are. You provide both the summary created by your system and one or more human-written 'reference' summaries. It then calculates various ROUGE scores, indicating how closely your system's output matches human quality, and outputs them in a CSV file. It's used by anyone developing or refining natural language processing models for summarization or translation.
220 stars. No commits in the last 6 months.
Use this if you need a reliable and standardized way to quantify the quality of summaries or translations produced by your machine learning models.
Not ideal if you're looking for a tool that generates summaries or translations itself, rather than evaluating them.
Stars
220
Forks
37
Language
Java
License
Apache-2.0
Category
Last pushed
Apr 09, 2020
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/kavgan/ROUGE-2.0"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
kenlimmj/rouge
A Javascript implementation of the Recall-Oriented Understudy for Gisting Evaluation (ROUGE)...
uoneway/KoBertSum
KoBertSum은 BertSum모델을 한국어 데이터에 적용할 수 있도록 수정한 한국어 요약 모델입니다.
udibr/headlines
Automatically generate headlines to short articles
bheinzerling/pyrouge
A Python wrapper for the ROUGE summarization evaluation package
xiongma/transformer-pointer-generator
A Abstractive Summarization Implementation with Transformer and Pointer-generator