https://github.com/zebbern/reconx
🕷️ | ReconX is a Live-Website Crawler made to gather critical information with an option to take a picture of each site crawled!
https://github.com/zebbern/reconx
crawler hacking information-gathering information-retrieval information-security livedata opsec osint osint-tool pentest python python-crawler search-engine security security-tools website website-crawler website-scraper website-security
Last synced: 6 months ago
JSON representation
🕷️ | ReconX is a Live-Website Crawler made to gather critical information with an option to take a picture of each site crawled!
- Host: GitHub
- URL: https://github.com/zebbern/reconx
- Owner: zebbern
- License: mit
- Created: 2025-02-05T08:26:51.000Z (11 months ago)
- Default Branch: main
- Last Pushed: 2025-02-20T16:36:54.000Z (10 months ago)
- Last Synced: 2025-04-13T16:51:09.370Z (9 months ago)
- Topics: crawler, hacking, information-gathering, information-retrieval, information-security, livedata, opsec, osint, osint-tool, pentest, python, python-crawler, search-engine, security, security-tools, website, website-crawler, website-scraper, website-security
- Language: Python
- Homepage:
- Size: 57.6 KB
- Stars: 3
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README



**ReconX is tool for crawling websites to gather OSINT data and identify potential vulnerabilities using customizable text-based patterns. It leverages a breadth-first search (BFS) strategy with multithreading for fast and efficient crawling.**
[Features](#features)
[Setup](#setup)
[Usage](#usage)
[Showcase](#showcase)
[Python 3.12+ Pip Fix](#python-312-pip-fix)
---
## Features
- **OSINT Gathering:** Extracts emails, phone numbers, subdomains, social links, and more.
- **Vulnerability Scanning:** Detects potential vulnerabilities by matching text patterns.
- **Customizable Patterns:** Uses external `.txt` files (in a `wordlists` folder) for scanning.
- **Concurrent Crawling:** Uses multithreading to speed up the crawling process.
- **Optional Screenshots:** Capture page screenshots using Selenium and Chrome WebDriver.
---
## Setup
1. **Clone or Download** `git clone https://github.com/zebbern/ReconX`
2. **Install requirements** `pip install -r requirements.txt` or `pip install requests beautifulsoup4 pyyaml selenium`
3. **Configuration:**
- (Optional) Change the `config.yaml` in the directory with settings you want or keep them default.
- (Optional) Adjust crawler settings like user agents, SSL verification, and screenshot options.
4. **Pattern Files:**
- Ensure the `wordlists` directory contains the following files:
- `query_params.txt`
- `api_wss.txt`
- `common_misconfig.txt`
- `injection_points.txt`
- `sensitive_data.txt`
- `exposed_files.txt`
- `js_vulnerabilities.txt`
- customise info in the .txt found in `/wordlists` folder.
## Usage
**Run the crawler by providing the base URL:**
python crawler.py -u https://example.com
### Terminal Options
- `-u`, `--url`
**Description:** Base URL to start crawling (required).
- `-t`, `--threads`
**Description:** Number of concurrent threads to use (default: 10).
- `-ws`
**Description:** Enable screenshot capture for the crawled pages.
**For all options on you can run**
#Example:
python crawler.py -u https://example.com -ws -t 10
**Explanation:**
- `-u https://example.com`: Specifies the base URL to crawl.
- `-ws`: Enables screenshots of fetched pages.
- `-t 10`: Uses 10 concurrent threads for faster crawling.
---
### Output
Upon completion, the tool generates two JSON files:
- **recon_data.json:** Contains collected OSINT data (e.g., URLs, emails, subdomains).
- **Vuln.json:** Contains details of detected vulnerabilities based on the pattern matches.
If screenshots are enabled (`-ws`), they will be saved in the directory specified by `config.yaml -> output_path: ""` (default: `webshots`).
---
## Showcase
`Command Used`: python ReconX.py -u https://monolittenil.no | Use `-ws` if you want to include pictures of each site its crawled can also edit `config.yaml` to crawl deeper.

## Python 3.12+ Pip Fix:
### Create and Activate a Virtual Environment
#### For Linux/macOS:
```
python3 -m venv venv && source venv/bin/activate
```
#### For Windows:
```
python -m venv venv && .\venv\Scripts\activate
```
> [!WARNING]
> These is intended for educational and ethical hacking purposes only. It should only be used to test systems you own or have explicit permission to test. Unauthorized use of third-party websites or systems without consent is illegal and unethical.