INNOVINATI/microwler
A micro-framework for asynchronous deep crawls and web scraping with Python
This tool helps developers systematically gather data from entire websites. You provide it with a website's starting URL and instructions on what data to look for on each page. It then navigates through all linked pages, collects the specified information, and gives you a structured dataset. It's designed for Python developers who need to perform deep web crawls.
Available on PyPI.
Use this if you are a Python developer and need to extract comprehensive data from every page of a specific website.
Not ideal if you only need to scrape a few pages or are not comfortable writing Python code for web scraping.
Stars
13
Forks
1
Language
Python
License
MIT
Category
Last pushed
Apr 04, 2026
Commits (30d)
0
Dependencies
13
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/perception/INNOVINATI/microwler"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
scrapy/scrapy
Scrapy, a fast high-level web crawling & scraping framework for Python.
Altimis/Scweet
A simple and unlimited twitter scraper : scrape tweets, likes, retweets, following, followers,...
lexiforest/curl_cffi
Python binding for curl-impersonate fork via cffi. A http client that can impersonate browser...
scrapinghub/spidermon
Scrapy Extension for monitoring spiders execution.
soxoj/maigret
🕵️♂️ Collect a dossier on a person by username from 3000+ sites