nkapila6/mcp-local-rag
"primitive" RAG-like web search model context protocol (MCP) server that runs locally. ✨ no APIs ✨
This tool helps large language models (LLMs) like Claude perform deep, up-to-date web research and answer questions using current information. It takes your natural language questions, searches across many web sources, extracts the most relevant details, and feeds them back to the LLM. Researchers, analysts, or anyone needing accurate, real-time information from an AI assistant would find this useful.
118 stars.
Use this if you want your AI assistant to conduct comprehensive, multi-source web searches and provide answers based on the very latest information, rather than just its training data.
Not ideal if you need to perform simple, quick fact-checks that don't require deep, multi-engine research or if you prefer manual web searching.
Stars
118
Forks
19
Language
Python
License
MIT
Category
Last pushed
Mar 09, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/mcp/nkapila6/mcp-local-rag"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related servers
shinpr/mcp-local-rag
Local-first RAG server for developers using MCP. Semantic + keyword search for code and...
AmberLee2427/nancy-brain
Nancy's RAG backend and HTTP API/MCP server connectors.
tac0de/knowledge-to-action-mcp
MCP server for Obsidian GraphRAG, agent-ready context, preview-only planning, and safe repo handoffs
Daniel-Barta/mcp-rag-server
Lightweight RAG server for the Model Context Protocol: ingest source code, docs, build a vector...
yumeiriowl/repo-graphrag-mcp
An MCP server that uses LightRAG and Tree-sitter to build a repository knowledge graph from code...