modulepackage
0.0.0-20241210145221-d61d8926601a
Repository: https://github.com/sohunn/status-crawler.git
Documentation: pkg.go.dev
# README
StatusCrawler
This is a simple tool used to detect dead links on a website and summarize their HTTP statuses in a clear table, written in Golang.
Features✨
- Supports and validates links using
http
andhttps
schemes. - Uses playwright to perform efficient web scraping.
- Leverages the power of go-routines with mutexes, wait groups and distributed locking mechanisms to increase performance and concurrency 🚀
- Clean summary in a tabular format.
How to use❓
-
Make sure you have the latest version of go installed.
-
Clone the repository using the following command:
git clone https://github.com/sohunn/status-crawler.git
- Install dependencies:
go mod tidy
- Make sure to install the browsers and OS dependencies:
go run github.com/playwright-community/playwright-go/cmd/playwright@latest install --with-deps
- From the root of the project:
go run ./ <URL>
Example
go run ./ "https://sohunn.me"
Building 🛠️
Check your Go env variables (GOOS
and GOPATH
) to make sure you are building the executable for the right platform. Once verified, run:
go build -o crawler.exe ./
Note: You can call your executable whatever you want. I have specified crawler
in the example
Once done, simply run the executable with the arguments like you normally would.
crawler.exe "https://sohunn.me"
# Functions
No description provided by the author
No description provided by the author
No description provided by the author
# Type aliases
No description provided by the author