Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/chandru-mc/web-scraping
This repository contains a Python-based web scraping project that extracts product data from Amazon. The project includes a Jupyter Notebook (Web_Scarping_amaz.ipynb) that demonstrates how to scrape product titles, prices, ratings, reviews, and availability using BeautifulSoup and requests. The scraped data is saved in a CSV file (amazon_data.csv)
https://github.com/chandru-mc/web-scraping
Last synced: 4 days ago
JSON representation
This repository contains a Python-based web scraping project that extracts product data from Amazon. The project includes a Jupyter Notebook (Web_Scarping_amaz.ipynb) that demonstrates how to scrape product titles, prices, ratings, reviews, and availability using BeautifulSoup and requests. The scraped data is saved in a CSV file (amazon_data.csv)
- Host: GitHub
- URL: https://github.com/chandru-mc/web-scraping
- Owner: CHANDRU-MC
- Created: 2024-08-27T17:23:28.000Z (3 months ago)
- Default Branch: main
- Last Pushed: 2024-08-27T17:26:39.000Z (3 months ago)
- Last Synced: 2024-08-27T19:16:46.989Z (3 months ago)
- Language: Jupyter Notebook
- Homepage: https://chandru-mc.github.io/web-scraping/
- Size: 6.84 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
## Overview
This project demonstrates web scraping techniques using Python. The primary goal is to scrape data from e-commerce websites and analyze it. This repository includes a Jupyter Notebook for scraping product information from Amazon.
## Project Details
The project includes the following components:this done in jupyter notebook
- **`Web_Scarping_amaz.ipynb`**: A Jupyter Notebook that performs web scraping on Amazon to extract product titles, prices, ratings, reviews, and availability.
- **`amazon_data.csv`**: The CSV file containing the scraped data.## Requirements
To run the web scraping notebook, you need to have the following Python packages installed:
- `requests`
- `beautifulsoup4`
- `pandas`
- `numpy`You can install the required packages using pip:
```bash
pip install requests beautifulsoup4 pandas numpy