scrapy crawl test -s JOBDIR=data -0 items.csv
watch -n 1 -d 'wc -l items.csv;du -h items.csv'
附数据 items.csv
用于爬取JD和github的代码,数据存储于mongodb。.
pip install scrapy-jingdong==1.2.1
scrapy crawl test -s JOBDIR=data -0 items.csv
watch -n 1 -d 'wc -l items.csv;du -h items.csv'
附数据 items.csv