pragmar/mcp-server-webcrawl

MCP server tailored to connecting web crawler data and archives

47
/ 100
Emerging

This project helps anyone working with large collections of website data or archives to search and retrieve specific information efficiently. It takes data from various web crawling tools like ArchiveBox or HTTrack, and lets you use advanced search queries to find exactly what you need. The output is filtered web content and analyses, which an AI client can then process. This tool is for web archivists, SEO specialists, digital marketers, or researchers who need to make sense of extensive crawled web data.

Use this if you need to perform detailed searches, audits, or analyses on large datasets collected by web crawlers, enabling an AI to efficiently process and understand the content.

Not ideal if you are looking for a web crawler itself, or if you only need to perform basic keyword searches on small, static website archives.

web-archiving SEO-auditing digital-research content-analysis data-retrieval
No Package No Dependents
Maintenance 6 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 18 / 25

How are scores calculated?

Stars

37

Forks

14

Language

Python

License

Category

web-scraping-mcp

Last pushed

Dec 08, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/mcp/pragmar/mcp-server-webcrawl"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.