panruotong/CAG
Implementation of Not All Contexts Are Equal: Teaching LLMs Credibility-aware Generation. Paper: https://arxiv.org/abs/2404.06809
This project helps evaluate how well large language models (LLMs) answer questions when provided with source documents that have varying levels of credibility. You input a question and several source documents, each marked with a credibility rating (high, medium, or low), and it assesses the quality of the LLM's answer. This is primarily useful for researchers and developers working with LLMs in question-answering systems.
No commits in the last 6 months.
Use this if you are a researcher or developer who needs to evaluate the performance of LLMs in generating accurate and reliable answers from potentially mixed-credibility sources.
Not ideal if you are looking for a ready-to-use application to answer your own questions without needing to assess or develop LLMs.
Stars
22
Forks
2
Language
Python
License
MIT
Category
Last pushed
Oct 22, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/rag/panruotong/CAG"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
denser-org/denser-retriever
An enterprise-grade AI retriever designed to streamline AI integration into your applications,...
rayliuca/T-Ragx
Enhancing Translation with RAG-Powered Large Language Models
neuml/rag
🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with...
NovaSearch-Team/RAG-Retrieval
Unify Efficient Fine-tuning of RAG Retrieval, including Embedding, ColBERT, ReRanker.
RulinShao/retrieval-scaling
Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".