Skip to content Skip to sidebar Skip to footer

Input/output For Scrapyd Instance Hosted On An Amazon EC2 Linux Instance

Recently I began working on building web scrapers using scrapy. Originally I had deployed my scrapy projects locally using scrapyd. The scrapy project I built relies on accessing d

Solution 1:

Is S3 an option? I'm asking because you're already using EC2. If that's the case, you could read/write from/to S3.

I'm a bit confused because you mentioned both CSV and JSON formats. If you're reading CSV, you could use CSVFeedSpider. Either way, you could also use boto to read from S3 in your spider's __init__ or start_requests method.

Regarding the output, this page explains how to use feed exports to write the output of a crawl to S3.

Relevant settings:


Post a Comment for "Input/output For Scrapyd Instance Hosted On An Amazon EC2 Linux Instance"