explore-de/rage4j
Evaluate your LLM based Java Apps
This tool helps Java developers evaluate the quality of their Language Model (LLM) based applications, especially those using Retrieval-Augmented Generation (RAG). It takes in your LLM's outputs, along with questions and expected answers, and provides metrics like correctness, relevance, and faithfulness. Java developers building or maintaining LLM applications will find this useful for testing and quality assurance.
Use this if you are a Java developer building LLM-powered applications and need to systematically test and quantify the accuracy and reliability of your model's responses.
Not ideal if you are not a Java developer or if you are looking for a no-code solution to evaluate LLM performance.
Stars
11
Forks
2
Language
Java
License
—
Category
Last pushed
Mar 02, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/explore-de/rage4j"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
amazon-science/auto-rag-eval
Code repo for the ICML 2024 paper "Automated Evaluation of Retrieval-Augmented Language Models...
ibm-self-serve-assets/JudgeIt-LLM-as-a-Judge
Automation Framework using LLM-as-a-judge to evaluate of Agentic AI, RAG, Text2SQL at scale;...
mit-ll-ai-technology/llm-sandbox
Large language model evaluation framework for logic and open-ended Q&A with a vareity of RAG and...
nl4opt/ORQA
[AAAI 2025] ORQA is a new QA benchmark designed to assess the reasoning capabilities of LLMs in...