From d8631ddf0544e3dc2b98b855fc95b42d5d8de3e8 Mon Sep 17 00:00:00 2001 From: Nick Ruest Date: Sun, 30 Apr 2023 20:05:45 -0400 Subject: [PATCH] Add crau. (#127) - Resolves #95 --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index c8f1f2e..bf7d554 100644 --- a/README.md +++ b/README.md @@ -66,6 +66,7 @@ This list of tools and software is intended to briefly describe some of the most * [Brozzler](https://github.com/internetarchive/brozzler) - A distributed web crawler (爬虫) that uses a real browser (Chrome or Chromium) to fetch pages and embedded urls and to extract links. *(Stable)* * [Cairn](https://github.com/wabarc/cairn) - A npm package and CLI tool for saving webpages. *(Stable)* * [Chronicler](https://github.com/CGamesPlay/chronicler) - Web browser with record and replay functionality. *(In Development)* +* [crau](https://github.com/turicas/crau) - crau is the way (most) Brazilians pronounce crawl, it's the easiest command-line tool for archiving the Web and playing archives: you just need a list of URLs. *(Stable)* * [Crawl](https://git.autistici.org/ale/crawl) - A simple web crawler in Golang. *(Stable)* * [crocoite](https://github.com/promyloph/crocoite) - Crawl websites using headless Google Chrome/Chromium and save resources, static DOM snapshot and page screenshots to WARC files. *(In Development)* * [DiskerNet](https://github.com/dosyago/DiskerNet) - A non-WARC-based tool which hooks into the Chrome browser and archives everything you browse making it available for offline replay. *(In Development)*