Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/arv-anshul/99acres-scrape

Scrape 99scrape.com with python using curl command and export data as CSV.
https://github.com/arv-anshul/99acres-scrape

99acres api asyncio curl-parser curler project python3 real-estate real-estate-website streamlit web-scraping

Last synced: 26 days ago
JSON representation

Scrape 99scrape.com with python using curl command and export data as CSV.

Awesome Lists containing this project

README

        

# Scrape 99Acres

> \[!CAUTION\]
>
> **Message from Author:** I ([Anshul Raj Verma](https://github.com/arv-anshul)) am not
> able to scrape 99acres website using this streamlit app due to some unathorization
> issue.
>
> BTW, you can use and try to scrape the website and if you are able to scrape, please
> raise a issue to discuss the problem. Thanks!

> [!IMPORTANT]
>
> **DISCLAIMER:** This project is only for education purpose.

### Why this project?

I am working on a real estate project. So I need real-time data for that so I write a program to gather the data for this project and convert it into a web-app using streamlit.

Check out my Data Analysis on the scrapped data [here](https://github.com/arv-anshul/campusx-project-notebooks).

### Datasets

I've scrape and uploaded some cities data in my Kaggle profile you can download that
and practice. It contains **more than 40k+** properties details from different Indian cities.

Dataset on Kaggle:
[**Indian Real Estate - 99acres.com**](https://www.kaggle.com/datasets/arvanshul/gurgaon-real-estate-99acres-com)

### Documentation

Documentation for this project is available in [🗒️ wiki](https://github.com/arv-anshul/99acres-scrape/wiki) section.

In order to see the EDA process on the data fetched by this app goto [📁 campusx-project-notebooks](https://github.com/arv-anshul/campusx-project-notebooks).

### Techs

1. Python>=3.11
2. Asynchronous Programming
3. Streamlit
4. Http Requests
5. Pydantic

### Setup

1. Install all the required packages.

```sh
pip install -r requirements.txt
```

### Usage

1. Run the streamlit app.

```sh
streamlit run app.py
```

2. Goto URL `http://localhost:8501/`.
3. Fill the form: **Select the city** which you want to scrape and submit.
4. After some backend processing; **a download button** will appear, click it to download the scrapped data.

### Disclaimer

As I wanted to scrape the data from 99Acres website. I am ensuring that I am not performing any illegal activity using this data. I used this data in my project to build some ML model and perform some data analysis.