A Web Scraping application written in Go that helps you find all dead links in your website
- Concurrently scan all pages of a website for dead links
- Handle dynamic content scraping with headless browsers
- Customizable scan depth
- Customizable concurrency level
- Export the results to a CSV file
- Export the results to a JSON file
- Clone the repository
git clone https://github.com/yingtu35/dead-link-hunter.git
cd dead-link-hunter- Run the server directly or build the binary
go run cmd/app/main.go --url yourwebsite.comor
go build -o dead-link-hunter cmd/app/main.go
./dead-link-hunter --url yourwebsite.com| Flag | Description | Default | Required |
|---|---|---|---|
--url |
Website URL to scan for dead links | - | Yes |
--static |
Enable static mode (faster but doesn't render JavaScript) | false |
No |
--export |
Export format (csv or json) |
- | No |
--filename |
Name of the export file (without extension) | result |
No |
--maxDepth |
Maximum crawl depth from starting URL | 5 | No |
--maxConcurrency |
Maximum number of concurrent requests | 20 | No |
--timeout |
Request timeout in seconds | 10 | No |
# Basic usage with default settings
./dead-link-hunter --url example.com
# Static scan with custom concurrency and export to CSV
./dead-link-hunter --url example.com --static --maxConcurrency 20 --export csv
# Deep scan with longer timeout and JSON export
./dead-link-hunter --url example.com --maxDepth 10 --timeout 20 --export json --filename deep-scan- Support for JavaScript rendering with headless browsers
- Add support for custom scan depth
- Add support for custom concurrency level
- Add support for exporting results
This project is licensed under the MIT License - see the LICENSE file for details.