Skip to content

adilsoncarvalho/barateza-nfcrawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

barateza/nfcrawler waffle.io dashboard

Docker

We use Docker as our container engine and it should run smoothly on most systems.

Building the image locally

docker build . -t barateza-nfcrawler
docker-compose build
docker-compose up --build

Invoking the shell on the container

docker run --rm -ti barateza-nfcrawler sh
docker-compose run --rm scrapy sh

Removing the image

docker rmi barateza-nfcrawler
docker-compose down

Running the crawler

docker run --rm -ti barateza-nfcrawler

Options as env vars

Pass them to the container using the -e/--env option. It is also possible to pass them using a --env-file option.

  • CACHEDIR location for the items file (defaults to /scrapy/nfcrawler/.scrapy/httpcache)
  • DATAFILE location for the items file (defaults to /scrapy/nfcrawler/.scrapy/data/items.json)
  • LOGFILE location for the log file (defaults to /scrapy/nfcrawler/.scrapy/log/logfile.log)
  • START_URL location of the page to get scraped. It can be also a file://

About

Crawler to get data from the NF-e and NFC-e

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published