antgroup/ravig-bench
Official implementation of "RAViG-Bench: A Benchmark for Retrieval-Augmented Visually-rich Generation with Multi-modal Automated Evaluation"
This tool helps you assess the quality of automatically generated web content, specifically HTML and CSS code, that incorporates visual elements and information retrieved from other documents. You feed in the generated web page code, and it provides detailed reports on whether the code works correctly, if the visual design looks good, and if the content is accurate and complete. It's ideal for developers or researchers who are building and testing systems that create web pages with rich visual layouts based on retrieved information.
Use this if you need to rigorously evaluate the functional correctness, visual appeal, and informational accuracy of automatically generated HTML/CSS web content.
Not ideal if you're manually creating web pages or only need to validate simple HTML without visual design or content quality checks.
Stars
10
Forks
—
Language
Python
License
Apache-2.0
Category
Last pushed
Jan 29, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/antgroup/ravig-bench"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
vectara/open-rag-eval
RAG evaluation without the need for "golden answers"
DocAILab/XRAG
XRAG: eXamining the Core - Benchmarking Foundational Component Modules in Advanced...
HZYAI/RagScore
⚡️ The "1-Minute RAG Audit" — Generate QA datasets & evaluate RAG systems in Colab, Jupyter, or...
AIAnytime/rag-evaluator
A library for evaluating Retrieval-Augmented Generation (RAG) systems (The traditional ways).
microsoft/benchmark-qed
Automated benchmarking of Retrieval-Augmented Generation (RAG) systems