KadekM/scrawler

Scala web crawling and scraping using fs2 streams

35
/ 100
Emerging

This project helps developers gather data from websites by defining how to navigate web pages and what information to extract. You provide a starting URL and rules for which links to follow and what data points (like text or URLs) to collect. It then outputs the extracted data, which can be further processed or stored.

No commits in the last 6 months.

Use this if you are a Scala developer building an application that needs to automatically collect specific content or follow links from various websites in a structured and efficient way.

Not ideal if you need a no-code solution or a tool with a graphical user interface for web scraping, as this is a library for Scala programmers.

web-scraping data-collection data-extraction content-mining site-crawling
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 13 / 25

How are scores calculated?

Stars

16

Forks

3

Language

HTML

License

MIT

Category

scraper

Last pushed

Aug 29, 2017

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/perception/KadekM/scrawler"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.