knowledge-to-action-mcp and mcp-rag-server

These two projects appear to be competing implementations of an MCP server designed for Retrieval Augmented Generation, both aiming to provide context to LLMs from user-provided documents.

mcp-rag-server
38
Emerging
Maintenance 10/25
Adoption 3/25
Maturity 20/25
Community 12/25
Maintenance 2/25
Adoption 7/25
Maturity 16/25
Community 13/25
Stars: 3
Forks: 1
Downloads:
Commits (30d): 0
Language: TypeScript
License: MIT
Stars: 25
Forks: 4
Downloads:
Commits (30d): 0
Language: TypeScript
License: MIT
No risk flags
Stale 6m No Package No Dependents

About knowledge-to-action-mcp

tac0de/knowledge-to-action-mcp

MCP server for Obsidian GraphRAG, agent-ready context, preview-only planning, and safe repo handoffs

Combines graph-aware note retrieval with optional embeddings-based semantic reranking to surface contextual neighbors, then structures results into agent-ready packets containing briefs, risks, and repo file hints. Implements preview-only action planning and bounded workspace inspection (ripgrep, git status) without exposing shell access, integrating via stdio with Claude, VS Code, and Cursor.

About mcp-rag-server

kwanLeeFrmVi/mcp-rag-server

mcp-rag-server is a Model Context Protocol (MCP) server that enables Retrieval Augmented Generation (RAG) capabilities. It empowers Large Language Models (LLMs) to answer questions based on your document content by indexing and retrieving relevant information efficiently.

This is a tool for developers who integrate large language models (LLMs) into applications. It takes your collection of documents, like text files or markdown, and turns them into a searchable index. This index then helps your LLM provide more accurate and context-aware answers based on your specific content, rather than just its general training data.

LLM-integration developer-tool information-retrieval contextual-AI AI-application-development

Scores updated daily from GitHub, PyPI, and npm data. How scores work