joaobenedetmachado/scrapit

A (really) easy way to web scrape

53
/ 100
Established

Defines scraping targets declaratively in YAML—selectors, transforms, validation, and output formats—eliminating the need to write Python code for new sources. Supports five fetch backends (BeautifulSoup, Playwright for JavaScript, httpx async, GraphQL, Bright Data) with 28+ field transforms, pagination, spider discovery, and parallel crawling. Outputs to eight formats (JSON, CSV, SQLite, MongoDB, PostgreSQL, Excel, Google Sheets, Parquet) with optional webhooks, change detection, Redis caching, and a built-in web dashboard.

No Package No Dependents
Maintenance 13 / 25
Adoption 8 / 25
Maturity 13 / 25
Community 19 / 25

How are scores calculated?

Stars

56

Forks

21

Language

Python

License

MIT

Last pushed

Mar 12, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/joaobenedetmachado/scrapit"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.