Categorygithub.com/Darkknight131714/go-web-crawler
modulepackage
0.0.0-20231023122804-1e198046147b
Repository: https://github.com/darkknight131714/go-web-crawler.git
Documentation: pkg.go.dev

# README

go-web-crawler

Hello there guys, This is a project for making a web crawler in go with some additional functionality like caching and other things.

Current Things built:

  1. Crawler: Done
  2. Caching: Done
  3. Retrying: Done(Will retry once regardless of error type in visiting. Can optimize here.)
  4. Priority Customers: Done(Paid Customers will go first on the basis of their timestamp. When all paid customers are done, unpaid customers will go on the basis of their timestamp.)
  5. Concurrency: Not Done
  6. Admin Access : Not Done

To use this:

Run : go get to fetch all the dependencies

Run : go run main.go to run the main file

This will create a localhost server which is listening on port 8080.

The only endpoint it serves currently is /crawl which is a get request.

This /crawl expects a url as a basepoint. This is a sample body for your http request to this server.

{

"Url":"<Your URL>"

}

# Structs

No description provided by the author
No description provided by the author
No description provided by the author