XRAG and RAG-evaluation-harnesses

XRAG
53
Established
Maintenance 10/25
Adoption 10/25
Maturity 16/25
Community 17/25
Maintenance 2/25
Adoption 6/25
Maturity 16/25
Community 11/25
Stars: 120
Forks: 18
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
Stars: 23
Forks: 3
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No Package No Dependents
Stale 6m No Package No Dependents

About XRAG

DocAILab/XRAG

XRAG: eXamining the Core - Benchmarking Foundational Component Modules in Advanced Retrieval-Augmented Generation

This project helps developers and researchers evaluate different components of Retrieval-Augmented Generation (RAG) systems. It takes various RAG configurations, such as different retrievers, embeddings, and Large Language Models, and outputs performance metrics and visualizations. The primary users are AI/ML engineers and researchers building or optimizing RAG applications.

RAG evaluation LLM benchmarking NLP research AI engineering Information retrieval

About RAG-evaluation-harnesses

RulinShao/RAG-evaluation-harnesses

An evaluation suite for Retrieval-Augmented Generation (RAG).

This project helps evaluate how well your Retrieval-Augmented Generation (RAG) system performs on various question-answering tasks. You provide your RAG model's retrieved documents and the questions, and it outputs performance scores. This tool is for researchers, developers, or MLOps engineers who are building and fine-tuning RAG systems and need to rigorously benchmark their effectiveness.

RAG-evaluation LLM-benchmarking NLP-research AI-model-testing information-retrieval

Scores updated daily from GitHub, PyPI, and npm data. How scores work