mcp-omnisearch and wet-mcp
The MCP server for unified access to multiple search engines and AI tools is a complementary service for the MCP server providing content extraction and documentation indexing, as the former can enhance the latter's capabilities by feeding it more diverse and intelligent search results and processed content.
About mcp-omnisearch
spences10/mcp-omnisearch
🔍 A Model Context Protocol (MCP) server providing unified access to multiple search engines (Tavily, Brave, Kagi), AI tools (Perplexity, FastGPT), and content processing services (Jina AI, Kagi). Combines search, AI responses, content processing, and enhancement features through a single interface.
This tool helps researchers, analysts, and content creators gather information efficiently by combining multiple search engines, AI answer tools, and web content processors into a single interface. You input a search query or a URL, and it provides web search results, AI-generated answers with citations, GitHub content, or extracted and summarized web page content. It's ideal for anyone needing comprehensive, multi-faceted online research.
About wet-mcp
n24q02m/wet-mcp
MCP server for web search, content extraction, and documentation indexing
Provides embedded metasearch (SearXNG) with semantic reranking and query expansion, plus specialized academic research across Google Scholar, arXiv, and PubMed. Features local full-text documentation indexing with HyDE-enhanced retrieval, batch content extraction from up to 50 URLs, and multimodal analysis—all with zero-config local embeddings (Qwen3) or optional cloud providers. Integrates as an MCP server with Claude, Gemini, and Codex via stdio transport, with automatic setup and encrypted credential storage.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work