Categorygithub.com/DAtek/grawler
modulepackage
0.3.0
Repository: https://github.com/datek/grawler.git
Documentation: pkg.go.dev

# README

codecov Go Report Card

Grawler

Simple and performant web crawler in Go

How it works

The crawler is using 2 types of workers:

  • Page loaders
  • Page analyzers

Page loaders are consuming the remaining URL channel and are downloading pages from the internet and putting them into a cache, also putting the downloaded page's URL into the downloaded URL channel.

Page analyzers are consuming the downloaded URL channel and reading the page's content from the cache, then analyzing the content, extracting additional URLs and the wanted model (if possible). The extracted new URLs are being put into the remaining URL channel, the found model in the result channel.

The whole process is being started with putting the starting URL into the remaining URL channel.

The number of Page loaders and Page analyzers are configurable.

Your possibilities are endless: you can implement your own cache, page loader and analyzer, the mocks and interfaces in the source will help you.

For guidance, please have a look at crawler_test.go.

The gopher was made with the Gopher Konstructor

# Packages

No description provided by the author
No description provided by the author

# Functions

No description provided by the author

# Structs

No description provided by the author
No description provided by the author
No description provided by the author
No description provided by the author

# Interfaces

No description provided by the author
No description provided by the author

# Type aliases

No description provided by the author