pyrouge and rouge

These are competitors offering alternative implementations of the same ROUGE metric family for evaluating summarization quality, with the first being a wrapper around the canonical ROUGE package while the second is a standalone implementation.

pyrouge
49
Emerging
rouge
35
Emerging
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 23/25
Maintenance 0/25
Adoption 6/25
Maturity 16/25
Community 13/25
Stars: 249
Forks: 72
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stars: 24
Forks: 4
Downloads:
Commits (30d): 0
Language: Perl
License: MIT
Stale 6m No Package No Dependents
Stale 6m No Package No Dependents

About pyrouge

bheinzerling/pyrouge

A Python wrapper for the ROUGE summarization evaluation package

This tool helps researchers and developers working on text summarization evaluate the quality of their automatically generated summaries. It takes your plain text summaries and corresponding 'gold standard' reference summaries, then processes them to produce standardized ROUGE scores. Anyone building or comparing different text summarization models would use this to quantify performance.

natural-language-processing text-summarization academic-research model-evaluation content-analysis

About rouge

neural-dialogue-metrics/rouge

An implementation of ROUGE family metrics for automatic summarization.

This tool helps researchers and developers working with natural language processing evaluate the quality of automatically generated summaries or translations. You input two pieces of text: a reference (the 'correct' version) and a candidate (the machine-generated version). It outputs scores (recall, precision, and F-measure) that indicate how well the candidate text matches the reference.

natural-language-processing text-summarization machine-translation nlp-evaluation computational-linguistics

Scores updated daily from GitHub, PyPI, and npm data. How scores work