Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/wangyihang/subdomain-crawler

A program for collecting subdomains of a list of given second-level domains (SLD)
https://github.com/wangyihang/subdomain-crawler

dns go penetration subdomain subdomain-enumeration subdomain-finder subdomain-scanner

Last synced: 2 months ago
JSON representation

A program for collecting subdomains of a list of given second-level domains (SLD)

Awesome Lists containing this project

README

        

# Subdomain Crawler

The program aims to help you collect subdomains of a list of given second-level domains (SLD).

![](assets/demo.gif)

## Installation

* Option 1: Download from [GitHub Releases](https://github.com/WangYihang/Subdomain-Crawler/releases/latest) directly (Recommended)
* Option 2: Go Install

```bash
$ go install github.com/WangYihang/Subdomain-Crawler/cmd/subdomain-crawler@latest
```

## Usage

1. Edit input file `input.txt`

```bash
$ head input.txt
tsinghua.edu.cn
pku.edu.cn
fudan.edu.cn
sjtu.edu.cn
zju.edu.cn
```

2. Run the program

```bash
$ subdomain-crawler --help
Usage:
subdomain-crawler [OPTIONS]

Application Options:
-i, --input-file= The input file (default: input.txt)
-o, --output-folder= The output folder (default: output)
-t, --timeout= Timeout of each HTTP request (in seconds) (default: 4)
-n, --num-workers= Number of workers (default: 32)
-d, --debug Enable debug mode
-v, --version Version

Help Options:
-h, --help Show this help message

$ subdomain-crawler
```

3. Check out the result in `output/` folder.

```bash
$ head output/*
```