pythonrouge and rouge-metric

These two tools are competitors, as both provide Python wrappers for evaluating summarization quality using ROUGE, making them alternative choices for the same task within the "transformer-based-summarization" ecosystem.

pythonrouge
47
Emerging
rouge-metric
31
Emerging
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 21/25
Maintenance 0/25
Adoption 6/25
Maturity 25/25
Community 0/25
Stars: 162
Forks: 34
Downloads:
Commits (30d): 0
Language: Perl
License: MIT
Stars: 21
Forks:
Downloads:
Commits (30d): 0
Language: Perl
License: MIT
Stale 6m No Package No Dependents
Stale 6m

About pythonrouge

tagucci/pythonrouge

Python wrapper for evaluating summarization quality by ROUGE package

This tool helps researchers and developers working on text summarization evaluate how good their automatically generated summaries are. You provide your system's summaries and a set of human-written reference summaries, and it calculates various ROUGE scores like ROUGE-1, ROUGE-2, and ROUGE-SU4. This is for anyone building or comparing automated summarization models.

text summarization natural language processing computational linguistics content evaluation

About rouge-metric

li-plus/rouge-metric

A Python wrapper of the official ROUGE-1.5.5.pl script and a re-implementation of full ROUGE metrics.

This tool helps researchers and developers automatically assess the quality of text summarization models. You input the summaries generated by your model and a set of human-written reference summaries. It then outputs various ROUGE scores, which are metrics indicating how well your generated summaries match the reference summaries. Anyone working with natural language processing, particularly in text generation or summarization, would find this useful.

text-summarization natural-language-processing NLP-evaluation content-analysis AI-model-assessment

Scores updated daily from GitHub, PyPI, and npm data. How scores work