Skip to content

A Pipeline for extracting data from Carlist.my and load to ElasticSearch

Notifications You must be signed in to change notification settings

natalieconan/crawly

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 

Repository files navigation

CRAWLY

Scrapy Scripts for scraping data

Run Commands

  • Clone Project and go to crawly/

    $ git clone git@github.com:natalieconan/crawly.git
    $ cd crawly
    
  • (Optional) To install pipenv with Homebrew:

    $ brew install pipenv
    
  • Activate Python Virtual Env using Pipenv and install packages

    $ pipenv shell
    $ pipenv install
    
  • Finally, run spider for crawling

    $ scrapy crawl ${spider_name}
    

    In this case spider_name = carlist, so run this command to start crawling:

    $ scrapy crawl carlist
    

About

A Pipeline for extracting data from Carlist.my and load to ElasticSearch

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages