When used with scrapy, use it with Scrapy Scheduler provided by scrapy-frontera. Scrapy scheduler provided
by Frontera is not supported. scrapy-frontera
is a scrapy scheduler which allows to use frontera backends,
like the present one, with scrapy projects.
See specific usage instructions at module and class docstrings at backend.py. Some examples of usage can be seen in the scrapy-frontera README.
A complete tutorial for using hcf-backend
with ScrapyCloud workflows is available at
shub-workflow Tutorial: Managing Hubstorage Crawl Frontiers. shub-workflow
is a framework for
defining workflows of spiders and scripts running over ScrapyCloud. This is a strongly recommended lecture, because it documents the integration of different tools
which together provide the best benefit.
Package also provides a convenient command line tool for hubstorage frontier handling and manipulation: hcfpal.py. It supports dumping, count, deletion, moving, listing, etc. See command line help for usage.
Another provided tool is crawlmanager.py. It facilitates the
scheduling of consumer spider jobs. Examples of usage are also available in the already mentioned shub-workflow
Tutorial.
pip install hcf-backend
For hcf-backend developers, Pipfile files are provided for a development environment.
Run:
$ pipenv install --dev
$ pipenv shell
$ cp .envtemplate .env
and edit .env accordingly