Pycon.cz 2019 scrapy workshop
🐹 🐸 This workshop should learn basic work with a Scrapy framework.
- Scrapy
- scrapy official documentation - good to read all :)
- scrapinghub sample projects Github - ❗ spider and test web scraped during this workshop is one of the sample projects from this repo
- scrapinghub organisation Github - many extremely useful projects, which extends scrapy in some way: Spidermon (monitoring), Splash (javascript engine), etc.
- scrapinghub blog - actualities, examples, best-prectices, etc.
- XPath
- XPath tutorial
- XPath concise
- XPath tutorial czech - in czech language
Git is a version control software that records changes to a file or set of files. Git is especially helpful for software developers as it allows changes to be tracked (including who and when) when working on a project.
To download Git, go to the following link and choose the correct version for your operating system: https://git-scm.com/downloads.
Docker is a software to develope and run containerized applications.
Application runs isolated in a container and is independent on the underlying operating system. The same container
can be run on any OS, where docker is installed. For a purpose of this tutorial I prepared container containing Scrapy
installed on a top of Ubuntu OS, to prevent possible problems with direct Scrapy installation on various operating systems.
Docker installation for Ubuntu
Docker installation for Windows (Windows 10 64bit: Pro, Enterprise or Education)
Docker installation for Mac
Scrapy installation guide - For a brave one, who wants to try installation directly to a OS. There are some python libraries, which themselves depend on non-Python packages that might require additional installation steps depending on your platform.
Scrapy installation quide - Windows - In a case you are not able to run docker on your Windows, install Scrapy directly.
-
Clone this repository. From your console.
git clone https://github.com/itsx/scrapy-workshop-pycon2019.git
-
Change into the newly created repository:
cd scrapy-workshop-pycon2019
-
From this directory run docker-compose command to start the container with installed scrapy:
sudo docker-compose up -d
(For windows users don't use sudo)
Now you can check that docker container is running with the command:
sudo docker-compose ps
-
Finally attach to a running container, and check version of installed scrapy:
sudo docker exec -it scrapy bash scrapy --version