https://github.com/scrapingant/alibaba_scraper
Alibaba scraper with using of rotating proxies and headless Chrome from ScrapingAnt
https://github.com/scrapingant/alibaba_scraper
alibaba-scraper datamining price-scraper price-scraping python scraper scraping scraping-api scraping-data scraping-tool scraping-web scraping-websites web-crawler web-crawler-python web-crawling
Last synced: 5 months ago
JSON representation
Alibaba scraper with using of rotating proxies and headless Chrome from ScrapingAnt
- Host: GitHub
- URL: https://github.com/scrapingant/alibaba_scraper
- Owner: ScrapingAnt
- Created: 2020-05-03T12:13:51.000Z (about 5 years ago)
- Default Branch: master
- Last Pushed: 2023-07-05T20:48:37.000Z (almost 2 years ago)
- Last Synced: 2023-07-06T08:06:38.389Z (almost 2 years ago)
- Topics: alibaba-scraper, datamining, price-scraper, price-scraping, python, scraper, scraping, scraping-api, scraping-data, scraping-tool, scraping-web, scraping-websites, web-crawler, web-crawler-python, web-crawling
- Language: Python
- Homepage: https://scrapingant.com
- Size: 151 KB
- Stars: 10
- Watchers: 4
- Forks: 3
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Alibaba parser using scrapingant.com
This project shows how to use ScrapingAnt scraping service to load public data from alibaba.ScrapingAnt takes away all the messy work necessary to set up a browser and proxies for crawling. So you can just focus on your data.
## Usage
To run this code you need RapidApi key. Just go to ScrapingAnt page on Rapidapi, and click "Subscribe to Test" button. After that you have to select plan(there is a free one including 100 requests). After that you can find you api key in "X-RapidAPI-Key" field on endpoints page.
#### With Docker
```shell script
docker build -t alibaba_scraper . && docker run -it -v ${PWD}/data:/app/data alibaba_scraper adidas --rapidapi_key
```#### Without Docker
This code was written for python 3.7+
```shell script
git clone https://github.com/ScrapingAnt/alibaba_scraper.git
cd alibaba_scraper
python3 -m venv .env
.env/bin/pip install -r requirements.txt
.env/bin/python main.py --help
.env/bin/python main.py adidas --rapidapi_key
```
#### Available params
```
.env/bin/python python main.py --helpUsage: main.py [OPTIONS] SEARCH_STRING
Options:
--rapidapi_key TEXT Api key from https://rapidapi.com/okami4kak/api/scrapingant [required]
--pages INTEGER Number of search pages to parse
--country [ae|br|cn|de|es|fr|gb|hk|in|it|il|jp|nl|ru|sa|us]
Country of proxies location
--help Show this message and exit.
```
#### Sample output:
Output is saved to data/ directory in csv format.
