RAG-evaluation-harnesses and RAG-Evaluator
Maintenance
2/25
Adoption
6/25
Maturity
16/25
Community
11/25
Maintenance
0/25
Adoption
3/25
Maturity
16/25
Community
14/25
Stars: 23
Forks: 3
Downloads: —
Commits (30d): 0
Language: Python
License: MIT
Stars: 4
Forks: 3
Downloads: —
Commits (30d): 0
Language: Python
License: MIT
Stale 6m
No Package
No Dependents
Stale 6m
No Package
No Dependents
About RAG-evaluation-harnesses
RulinShao/RAG-evaluation-harnesses
An evaluation suite for Retrieval-Augmented Generation (RAG).
This project helps evaluate how well your Retrieval-Augmented Generation (RAG) system performs on various question-answering tasks. You provide your RAG model's retrieved documents and the questions, and it outputs performance scores. This tool is for researchers, developers, or MLOps engineers who are building and fine-tuning RAG systems and need to rigorously benchmark their effectiveness.
RAG-evaluation
LLM-benchmarking
NLP-research
AI-model-testing
information-retrieval
About RAG-Evaluator
GURPREETKAURJETHRA/RAG-Evaluator
A library for evaluating Retrieval-Augmented Generation (RAG) systems
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work