Skip to content

bp72/crwl

Repository files navigation

Crawler

crwl is an open source web crawler in Golang which allows you to traverse entire site. Using it, you can scan, benchmark and validate your site, for example evaluate connected component or internal pagerank

Motivation

I faced problem to crawl site as-is for various reason: create set site structure as graph, validate it, benchmark.

Get Started

Clone repo

git clone git@github.com:bp72/crwl.git

Build

make build

Run

bin/crwl -domain example.com -use-internal-cache -max-depth 3 -max-workers 5

Crawler arch

alt text

Web Crawler Features

  • Start from the root domain and crawl the web pages with a specified depth.
  • Save the pages
  • Support logging and statsd metrics

TODO Features

  • Add WebUI to control and manage crawler
  • Add Crawl delay support per domain
  • Add Data storage interface to support FS, ClickHouse, RDB
  • Add logic to respect robots.txt
  • Add Grafana dashboard to repo
  • Add docker-compose to setup and run crawler with external service dependencies
  • Add condition to save page content to storage, for example keyword or url pattern

Options

Benchmark/Test mode

Sometime you just need to traverse your site without storing the content, just to check everything works fine or how far you can go. In this case you can use -do-not-store option, it disables content storing function :

bin/crwl -do-not-store

Setting up limits

Maximum crawls limitation Option allows to limit number of crawls with exact number, by default it's 100k pages to crawl

bin/crwl -max-crawl 1234

Maximum depth allows to set limitation on how deep crawler can go, by default it's 7

bin/crwl -max-depth 1

Maximum number of worker sets the limit of concurrent cralwers to run, by default it's 20

bin/crwl -max-workers 2

Run without any external service dependancy

Crawler can be run standalone (without other services), however this configuration has memory limitation, since it's maintaince urls queue and visitied url in memory.

bin/crwl -use-internal-cache

Metrics and logging

Crawler support statd metric publishing technique, to enable it:

bin/crwl -statsd-addr hostname:port

Roadmap

  • Define crawler arch
  • Implement initial crawler version
  • Add WebUI to control and manage crawler
  • Add Crawl delay support per domain
  • Add Data storage interface to support FS, ClickHouse, RDB
  • Respect robots.txt
  • Add Grafana dashboard to repo
  • Add docker-compose to setup and run crawler with external service dependencies

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published