suckit

Web scraper

A Rust-based web scraping tool that recursively visits and downloads websites to disk.

Suck the InTernet

GitHub

747 stars
10 watching
39 forks
Language: Rust
last commit: 8 months ago
Linked from 1 awesome list

hacktoberfestrustwebscraping

Backlinks from these awesome lists:

Related projects:

Repository Description Stars
spider-rs/spider A web crawler and scraper built on top of Rust, designed to extract data from the web in a flexible and configurable manner. 1,140
the-markup/blacklight-collector A tool for scraping website content and analyzing browser behavior 202
rust-scraper/scraper A Rust library for parsing and querying HTML documents using CSS selectors. 1,937
archiveteam/grab-site A web crawler designed to backup websites by recursively crawling and writing WARC files. 1,402
fimad/scalpel A web scraping library providing a declarative interface on top of an HTML parsing library to extract data from HTML pages 323
utkarshkukreti/select.rs A Rust library for extracting useful data from HTML documents 974
yhat/scrape A collection of utility functions and tools to simplify web scraping in Go. 1,513
spekulatius/phpscraper A web scraping utility for PHP that simplifies the process of extracting information from websites. 536
internetarchive/brozzler A distributed web crawler that fetches and extracts links from websites using a real browser. 671
vida-nyu/ache A web crawler designed to efficiently collect and prioritize relevant content from the web 454
jakopako/goskyr A tool to simplify web scraping of list-like structured data from web pages 35
archiveteam/wpull Downloads and crawls web pages, allowing for the archiving of websites. 556
tidyverse/rvest A package for extracting data from web pages using HTML parsing and CSS/XPath selectors. 1,492
felipecsl/wombat A Ruby-based web crawler and data extraction tool with an elegant DSL. 1,315
medialab/minet A command line tool and Python library for extracting data from various web sources. 286