Last updated:
0 purchases
osrqscrapy 0.0.7
os-rq-scrapy
A framework for Scrapy working with os-rq-pod and os-rq-hub to build "broad crawls" system.
As you know, Scrapy is a very popular python crawler framework. It is suit for "focused crawl", start from several URLs of specific sites, fetch html, extract and save "structured data" also with patternd links to crawl recursively. But for large scale, long time crawling especially "broad crawls", scrapy is incompetent. Basically, you have to decouple the whole crawling system into several sub-systems, high-performance full-featured distributed fetcher, task scheduler, html extractor, link database, data storage, proxy and a lot of auxiliary equipments. It will be more complex when your system is for multi-tenancy.
The os-rq-scrapy and os-rq-pod project are basic components to build "broad crawls" system. The core conceptions are very simple, os-rq-pod is multi-sites request queue have http api to recieve requests. os-rq-scrapy is fetcher, getting requests from os-rq-pod and crawl multi-sites concurrently. os-rq-hub can also be used to connect multi pod and scrapy instances to work simultaneously.
Requirements
Python 3.6+ (pypy3.6+)
Scrapy 2.0
extra requirements:
ujson, for json performance
Install
pip install os-rq-scrapy
Usage
Command line
rq-scrapy command enhance the basic scrapy command. When RQ_API is configured, the crawl subcommand will run on rq mode, get requests from rq.
Unit Tests
tox
License
MIT licensed.
For personal and professional use. You cannot resell or redistribute these repositories in their original state.
There are no reviews.