Skip to content

omkarcloud/dentalkart-scraper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

9 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

omkar

✨ DentalKart Scraper πŸ€–

πŸ’¦ Enjoy the Rain of DentalKart Products πŸ’¦

(Programming Language - Python 3)

dentalkart-scraper forks Repo stars dentalkart-scraper License issues

View

Open in Gitpod


Disclaimer for Dentalkart Scraper Project

By using Dentalkart Scraper, you agree to comply with all applicable local and international laws related to data scraping, copyright, and privacy. The developers of Dentalkart Scraper will not be held liable for any misuse of this software. It is the user's sole responsibility to ensure adherence to all relevant laws regarding data scraping, copyright, and privacy, and to use Dentalkart Scraper in an ethical and legal manner, in line with both local and international regulations.

We take concerns related to the Dentalkart Scraper Project very seriously. If you have any inquiries or issues, please contact Chetan Jain at chetan@omkar.cloud. We will take prompt and necessary action in response to your emails.


🌟 Scrape DentalKart Products! πŸ€–

This scraper is designed to help you download DentalKart Products.

πŸš€ Getting Started

1️⃣ Clone the Magic πŸ§™β€β™€οΈ:

git clone https://github.com/omkarcloud/dentalkart-scraper
cd dentalkart-scraper

2️⃣ Install Dependencies πŸ“¦:

python -m pip install -r requirements.txt

3️⃣ Let the Rain of DentalKart Products Begin 😎:

python main.py

Once the scraping process is complete, you can find your DentalKart Products in the output/finished.csv.

πŸ€” Questions

❓ Can I Interrupt the Scrape While It's Running?

Yes, you can. The scraper will resume from where it left off if you interrupt the process.

❓ I don't have Python, or I'm facing errors when setting up the scraper on my PC. How to solve it?

You can easily run the scraper in Gitpod, a browser-based development environment. Set it up in just 5 minutes by following these steps:

  1. Visit this link and sign up using your GitHub account.

    Screenshot (148)

  2. Once Signed Up, Open it in Gitpod.

    gp-continue

  3. In the terminal, run the following command to start scraping:

    python main.py
  4. Once the scraper has finished running, you can download the data from output folder.

    Screenshot (219)

Also, it's important to regularly interact with the Gitpod environment, such as clicking within it every 30 minutes, to keep the machine active and prevent automatic shutdown.

If you don't want to click every 30 minutes, then we encourage to install Python on PC and run the scraper locally.

Become one of our amazing stargazers by giving us a star ⭐ on GitHub!

It's just one click, but it means the world to me.

Stargazers for @omkarcloud/google-maps-scraper

Made with ❀️ using Botasaurus Web Scraping Framework