rag-evaluator and rageval

These two tools are competitors, as both provide libraries for evaluating Retrieval-Augmented Generation (RAG) systems.

rag-evaluator
52
Established
rageval
36
Emerging
Maintenance 0/25
Adoption 8/25
Maturity 25/25
Community 19/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 10/25
Stars: 42
Forks: 18
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stars: 170
Forks: 10
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
Stale 6m
Stale 6m No Package No Dependents

About rag-evaluator

AIAnytime/rag-evaluator

A library for evaluating Retrieval-Augmented Generation (RAG) systems (The traditional ways).

This tool helps you check the quality of answers generated by AI systems, especially those that combine information retrieval with text generation (RAG systems). You provide an AI's answer, the original question, and a perfect reference answer, and it tells you how good the AI's answer is. This is ideal for AI developers, researchers, and anyone building or testing conversational AI applications.

AI-development NLP-evaluation conversational-AI-testing content-generation-quality

About rageval

gomate-community/rageval

Evaluation tools for Retrieval-augmented Generation (RAG) methods.

This tool helps evaluate the performance of your Retrieval-Augmented Generation (RAG) systems. It takes the outputs from various stages of your RAG pipeline—like rewritten queries, retrieved documents, and generated answers—and provides comprehensive scores on how well your system is performing across aspects like answer correctness, factual consistency, and document relevance. It is designed for AI/ML engineers or researchers building and refining RAG-based applications.

AI-evaluation NLP-benchmarking Generative-AI-testing LLM-performance Information-retrieval-quality

Scores updated daily from GitHub, PyPI, and npm data. How scores work