Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/DiegoCaraballo/Email-extractor
The main functionality is to extract all the emails from one or several URLs - La funcionalidad principal es extraer todos los correos electrónicos de una o varias Url
https://github.com/DiegoCaraballo/Email-extractor
email email-extractor email-marketing emails extraction python scraper scrapers scraping scraping-websites scrapper scrapping scrapy scrapy-spider spyder stractor
Last synced: 3 months ago
JSON representation
The main functionality is to extract all the emails from one or several URLs - La funcionalidad principal es extraer todos los correos electrónicos de una o varias Url
- Host: GitHub
- URL: https://github.com/DiegoCaraballo/Email-extractor
- Owner: DiegoCaraballo
- Created: 2018-02-07T14:40:55.000Z (over 6 years ago)
- Default Branch: master
- Last Pushed: 2023-03-29T12:25:26.000Z (over 1 year ago)
- Last Synced: 2024-05-02T19:20:52.581Z (6 months ago)
- Topics: email, email-extractor, email-marketing, emails, extraction, python, scraper, scrapers, scraping, scraping-websites, scrapper, scrapping, scrapy, scrapy-spider, spyder, stractor
- Language: Python
- Homepage: https://whitemonkey.io
- Size: 111 KB
- Stars: 174
- Watchers: 15
- Forks: 71
- Open Issues: 10
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- awesome-hacking-lists - DiegoCaraballo/Email-extractor - The main functionality is to extract all the emails from one or several URLs - La funcionalidad principal es extraer todos los correos electrónicos de una o varias Url (Python)
README
# Add Feature: 13-07-2022
- You can save the mailing list in a .csv file# Fix: 13-09-2019
- Fix - The script was pasted when searching for phrases on Google.
- Add Requirements - pip install -r requirements.txt# Email Extractor Functions
## English
- (1) Extract emails from a single URL
- (2) Extract emails from a URL (Two Levels) - Search on the page and all its URLs
- (3) Do a Google search, save the Urls found and search the emails
- (4) Same as option 3 but with a list of keywords (TODO)
- (5) You can list the saved emails
- (6) You can save the mailing list in a .txt file
- (7) You can save the mailing list in a .csv file
- (8) Delete Emails from data base
- (9) Exit- The emails are stored in a Sqlite database ("Emails.db")
## Español
- (1) Extraer los correos de una única URL
- (2) Extraer los correos de una Url (Dos Niveles) - Busca sobre la página y todas sus URL
- (3) Hacer una busqueda en Google, guardar las Urls encontradas y buscar los correos en dichas Urls
- (4) Igual que la opción 3 pero con una lista de palabras (TODO)
- (5) Listar correos guardados
- (6) Se pueden guardar los correo en un archivo .txt
- (7) Se pueden guardar los correo en un archivo .csv
- (8) Eliminar correos de la base de datos
- (9) Salir- Todos los correos son guardados en una base de datos Sqlite ("Emails.db")
**Versión:** Python 3.x.
# Required modules - Modulos necesarios
pip install -r requirements.txt
[Extraer correos de paginas web con Python](http://www.pythondiario.com/2018/04/extraer-correos-electronicos-de-paginas.html)
## Docker
Docker and docker-compose are required.
In order to use docker follow below instructions:
### Installation
1. Get an .env file
```
cp .env.example .env
```2. Start docker container
```
docker-compose up -d --build
```### Usage
To execute the script and get the options menu:
```
docker exec -ti email-extractor python EmailExtractor.py
```To get the sqlite db with al e-mails:
```
docker cp email-extractor:Emails.db .
```To get the file saved, for instance, as "out":
```
ocker cp email-extractor:out.txt .
```