A high-performance web crawler written for broken-link detection and image downloading.
Overview: argiope is a fast, lightweight web crawler designed for broken-link detection, website archiving, and batch image downloading. It features intelligent HTML scanning, URL normalization, and sophisticated report generation. Built with zero external dependencies and compiled to a single static binary.
Key Features:
library.html landing page, nested index.html navigation, and per-folder reader.html viewersCommands:
check <url>: Crawl a website and generate a detailed broken-link report with timing statisticsimages <url>: Download all images from a website or manga chapters from MangaFox to an organized structurelibrary <dir>: Generate or regenerate HTML browser for existing image directoriesUsage Examples:
argiope check https://example.com --depth 5 — Check site for broken links up to 5 levels deepargiope images https://example.com/gallery -o ./images — Archive gallery images with HTML browserargiope images https://fanfox.net/manga/title --chapters 1-50 — Download manga chapters 1-50argiope check https://example.com --report report.html --report-format html — Generate HTML report for CI pipelinesargiope images https://example.com -o ./archive --parallel --depth 3 — Fast parallel crawling and downloadReport Features: Reports include detailed statistics: total URLs checked, OK count, broken count, error count, internal vs external split, and timing information (total time, average/min/max response times). HTML reports are self-contained with inline CSS and styled with status badges.
HTML Browser: The generated portable HTML browser features light/dark/system theme modes with localStorage-backed preferences, relative links for offline browsing, percent-encoded filenames for local file access, and thumbnail galleries with ordered prev/next navigation. Works with generic image collections and deep MangaFox chapter trees alike.
MangaFox Features:
--chapters N-M flag[manga-title]/[chapter]/[page].jpgPerformance & Reliability: Configurable request timeouts, delay between requests, and concurrent crawling options ensure efficient resource usage and reliable operation on rate-limited sites. Automatic redirect following, response size limiting, and detailed error reporting make it production-ready for CI pipelines and large-scale archiving. Perfect for website archiving, CI/CD integration, and offline browsing.
You are about to open
Do you wish to proceed?
Thank you for your report. Information you provided will help us investigate further.
There was an error while sending your report. Please try again later.
Snaps are applications packaged with all their dependencies to run on all popular Linux distributions from a single build. They update automatically and roll back gracefully.
Snaps are discoverable and installable from the Snap Store, an app store with an audience of millions.
Snap is available for CentOS 7.6+, and Red Hat Enterprise Linux 7.6+, from the Extra Packages for Enterprise Linux (EPEL) repository. The EPEL repository can be added to your system with the following command:
sudo yum install epel-release
Snap can now be installed as follows:
sudo yum install snapd
Once installed, the systemd unit that manages the main snap communication socket needs to be enabled:
sudo systemctl enable --now snapd.socket
To enable classic snap support, enter the following to create a symbolic link between /var/lib/snapd/snap and /snap:
sudo ln -s /var/lib/snapd/snap /snap
Either log out and back in again, or restart your system, to ensure snap’s paths are updated correctly.
To install argiope, simply use the following command:
sudo snap install argiope
Browse and find snaps from the convenience of your desktop using the snap store snap.
Interested to find out more about snaps? Want to publish your own application? Visit snapcraft.io now.