Skip to content

fuadhasanemon/Web-scraping

Repository files navigation

Web-scraping

#install scrapy pip install scrapy

#scrapy env conda install -c scrapinghub scrapy

#Creating a project Before you start scraping, you will have to set up a new Scrapy project. Enter a directory where you’d like to store your code and run:

scrapy startproject tutorial

This will create a tutorial directory with the following contents:

`
tutorial/ scrapy.cfg # deploy configuration file

tutorial/             # project's Python module, you'll import your code from here
    __init__.py

    items.py          # project items definition file

    middlewares.py    # project middlewares file

    pipelines.py      # project pipelines file

    settings.py       # project settings file

    spiders/          # a directory where you'll later put your spiders
        __init__.py 
        `

How to run our spider

To put our spider to work, go to the project’s top level directory and run:

scrapy crawl quotes

About

Web Scraping using Python

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published