7 Repositories
Rust scraping Libraries
A lightweight async Web crawler in Rust, optimized for concurrent scraping while respecting `robots.txt` rules.
🕷️ crawly A lightweight and efficient web crawler in Rust, optimized for concurrent scraping while respecting robots.txt rules. 🚀 Features Concurren
🌊 ~ seaward is a crawler which searches for links or a specified word in a website.
🌊 seaward Installation cargo install seaward On NetBSD a pre-compiled binary is available from the official repositories. To install it, simply run:
Web scraper written in Rust for extracting paid HTML templates for free.
HTML Template Scraper This tool provides a convenient solution by allowing you to access and download a wide range of HTML website templates, includin
A modern-looking, lightening fast, privacy respecting, secure, self-hostable meta search engine.
Websurfx A modern-looking, lightning-fast, privacy-respecting, secure meta search engine (pronounced as websurface or web-surface /wɛbˈsɜːrfəs/.) writ
Low effort scraping Python's pickle format in Rust. It is to complete pickle parsing as BeautifulSoup was to complete HTML parsing.
repugnant-pickle Because it is, isn't it? This is a Rust crate for dealing with the Python pickle format. It also has support for opening PyTorch file
A Rust library to extract useful data from HTML documents, suitable for web scraping.
select.rs A library to extract useful data from HTML documents, suitable for web scraping. NOTE: The following example only works in the upcoming rele
Rust crate for scraping URLs from HTML pages
url-scraper Rust crate for scraping URLs from HTML pages. Example extern crate url_scraper; use url_scraper::UrlScraper; fn main() { let director