An open API service indexing awesome lists of open source software.

https://github.com/parvvaresh/web-scraper-projects

This repository contains projects that include web scraping
https://github.com/parvvaresh/web-scraper-projects

bs4 python requests webscraping

Last synced: 6 months ago
JSON representation

This repository contains projects that include web scraping

Awesome Lists containing this project

README

          

# Web Scraper Projects

This repository contains a collection of web scraping projects aimed at various purposes. Whether you're interested in monitoring news websites, analyzing data, or fetching real-time data like currency exchange rates, this repository offers a range of tools to help you achieve your goals.

## Projects Overview

### 1. Nan (The country's needs)

#### Description
This project focuses on gathering and analyzing data related to the country's needs. It includes tools for scraping websites, extracting relevant information, and performing data analysis.

#### Files
- `data_analyze_v1_3.ipynb`: Jupyter Notebook containing data analysis scripts.
- `Data_nan.csv`: CSV file containing the scraped data.
- `links.txt`: Text file containing links used for scraping.

#### Source Code
- `src/get_page.py`: Python script for fetching web pages.
- `src/main.py`: Main script for running the scraping process.
- `src/save_link.py`: Script for saving scraped links.

### 2. News

#### Description
This project focuses on scraping news articles from specific sources. It provides tools to extract news articles from designated websites.

#### Source Code
- `src/fars_news.py`: Python script for scraping news articles from Fars News website.

### 3. Price - Dollar

#### Description
This project aims to fetch real-time data related to the dollar exchange rate. It provides tools to scrape currency exchange information from relevant websites.

#### Source Code
- `src/get_info.py`: Python script for fetching currency exchange information.
- `src/main.py`: Main script for running the scraping process.

## Getting Started

To get started with any of the projects, follow these steps:

1. Clone the repository to your local machine.
2. Navigate to the project directory you're interested in.
3. Install any necessary dependencies listed in the `requirements.txt` file.
4. Run the main script to start scraping data.

## Contribution

Contributions to this repository are welcome! Whether it's adding new scraping scripts, improving existing ones, or fixing bugs, feel free to submit pull requests.

## License

This repository is licensed under the [MIT License](LICENSE).

---