lm-rag-techniques and rag-evaluation

lm-rag-techniques
30
Emerging
rag-evaluation
29
Experimental
Maintenance 0/25
Adoption 2/25
Maturity 16/25
Community 12/25
Maintenance 0/25
Adoption 6/25
Maturity 8/25
Community 15/25
Stars: 2
Forks: 1
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: MIT
Stars: 17
Forks: 4
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License:
Stale 6m No Package No Dependents
No License Stale 6m No Package No Dependents

About lm-rag-techniques

NamaWho/lm-rag-techniques

Question-Answering (QA) system powered by Retrieval-Augmented Generation (RAG). The system leverages advanced methods such as Rank Fusion and Cascading Retrieval for optimized document retrieval and contextual QA generation.

About rag-evaluation

0xshre/rag-evaluation

A QA RAG system that uses a custom chromadb to retrieve relevant passages and then uses an LLM to generate the answer.

This project helps evaluate and improve question-answering systems built using Retrieval-Augmented Generation (RAG). You feed in documents and questions, and it generates answers while also providing a detailed report on how accurate and relevant the answers are. It's for data scientists and AI engineers who are developing or fine-tuning RAG-based chatbots or knowledge retrieval tools.

AI-development Natural-Language-Processing Knowledge-retrieval ML-evaluation Question-Answering-systems

Scores updated daily from GitHub, PyPI, and npm data. How scores work