my8100/scrapyd-cluster-on-heroku

Set up free and scalable Scrapyd cluster for distributed web-crawling with just a few clicks. DEMO :point_right:

49
/ 100
Emerging

This project helps data analysts and researchers set up a robust, scalable system for gathering information from websites. You provide a list of websites you want to collect data from, and the system automatically distributes the work, extracts the information you need, and stores it for your analysis. This is ideal for anyone who needs to collect large amounts of public data from multiple websites efficiently and reliably.

123 stars. No commits in the last 6 months.

Use this if you need to perform large-scale, automated web scraping across many sites and want a free, easily scalable setup.

Not ideal if your web scraping needs are small-scale or if you require persistent storage directly on the scraping servers themselves without an external database.

web-scraping data-collection market-intelligence competitive-analysis research-data
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 23 / 25

How are scores calculated?

Stars

123

Forks

81

Language

Python

License

GPL-3.0

Category

scraper

Last pushed

Apr 04, 2020

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/perception/my8100/scrapyd-cluster-on-heroku"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.