{"id":13603099,"url":"https://github.com/utkusen/urlhunter","last_synced_at":"2025-05-14T15:07:48.517Z","repository":{"id":40003593,"uuid":"314763206","full_name":"utkusen/urlhunter","owner":"utkusen","description":"a recon tool that allows searching on URLs that are exposed via shortener services","archived":false,"fork":false,"pushed_at":"2025-01-23T15:48:12.000Z","size":53,"stargazers_count":1580,"open_issues_count":0,"forks_count":109,"subscribers_count":63,"default_branch":"main","last_synced_at":"2025-04-13T21:33:49.213Z","etag":null,"topics":["bugbounty","intelligence","osint","recon","security"],"latest_commit_sha":null,"homepage":"","language":"Go","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/utkusen.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2020-11-21T08:05:57.000Z","updated_at":"2025-04-13T20:33:52.000Z","dependencies_parsed_at":"2022-07-10T15:33:12.687Z","dependency_job_id":"5ce91f25-30ff-4f6f-a66e-e242ca09a332","html_url":"https://github.com/utkusen/urlhunter","commit_stats":null,"previous_names":[],"tags_count":4,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/utkusen%2Furlhunter","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/utkusen%2Furlhunter/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/utkusen%2Furlhunter/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/utkusen%2Furlhunter/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/utkusen","download_url":"https://codeload.github.com/utkusen/urlhunter/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":254169744,"owners_count":22026214,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["bugbounty","intelligence","osint","recon","security"],"created_at":"2024-08-01T18:01:50.503Z","updated_at":"2025-05-14T15:07:48.478Z","avatar_url":"https://github.com/utkusen.png","language":"Go","readme":"```\n\t\t\t\to  \t  Utku Sen's\n\t\t\t\t \\_/\\o   \n\t\t\t\t( Oo)                    \\|/\n\t\t\t\t(_=-)  .===O-  ~~U~R~L~~ -O-\n\t\t\t\t/   \\_/U'        hunter  /|\\\n\t\t\t\t||  |_/\n\t\t\t\t\\\\  |    utkusen.com\n\t\t\t\t|K ||\ttwitter.com/utkusen_en\n```\n\nurlhunter is a recon tool that allows searching on URLs that are exposed via shortener services such as bit.ly and goo.gl. The project is written in Go.\n\n### How?\n\nA group named [URLTeam](https://archiveteam.org/index.php?title=URLTeam) (kudos to them) are brute forcing the URL shortener services and publishing matched results on a daily basis. urlhunter downloads their [collections](https://archive.org/details/UrlteamWebCrawls) and lets you analyze them. \n\n# Installation\n\n## From Binary\n\nYou can download the pre-built binaries from the [releases](https://github.com/utkusen/urlhunter/releases/latest) page and run. For example:\n\n`tar xzvf urlhunter_0.1.0_Linux_amd64.tar.gz`\n\n`./urlhunter --help`\n\n## From Source\n\n1) Install Go on your system\n\n2) Run: `go install github.com/utkusen/urlhunter@latest`\n\n**Note For The Windows Users:** urlhunter uses `XZ Utils` which is pre-installed on Linux and macOS systems. For Windows systems, you need to download it from [https://tukaani.org/xz/](https://tukaani.org/xz/)\n\n# Usage\n\nurlhunter requires 3 parameters to run: `-keywords`, `-date`. \n\nFor example: `urlhunter --keywords keywords.txt --date 2020-11-20`\n\n### --keywords\n\nYou need to specify the txt file that contains keywords to search on URLs. Keywords must be written line by line. You have three different ways to specify keywords:\n\n**Single Keyword:** urlhunter will search the given keyword as a substring. For example:\n\n`acme.com` keyword will both match `https://acme.com/blabla` and `https://another.com/?referrer=acme.com`\n\n**Multiple Keywords:** urlhunter will search the given keywords with an `AND` logic. Which means, a URL must include all the provided keywords. Keywords must be separated with `,` character. For example:\n\n`acme.com,admin` will match `https://acme.com/secret/adminpanel` but **_won't_** match `https://acme.com/somethingelse`\n\n**Regex Values:** urlhunter will search for the given regex value. In the keyword file, the line that contains a regular expression formula must start with `regex ` string. The format is: `regex REGEXFORMULA`. For example:\n\n`regex 1\\d{10}` will match `https://example.com/index.php?id=12938454312` but **_won't_** match `https://example.com/index.php?id=abc223`\n\n### --date\n\nurlhunter downloads the archive files of the given date(s). You have three different ways to specify the date:\n\n**Latest:** urlhunter will download the latest archive. `-date latest`\n\n**Single Date:** urlhunter will download the archive of the given date. Date format is YYYY-MM-DD. \n\nFor example: `-date 2020-11-20`\n\n**Year:** urlhunter will download the archive of the given year. Year format is YYYY.\n\nFor example: `-date 2024`\n\n**Date Range:** urlhunter will download all the archives between given start and end dates. \n\nFor example: `-date 2020-11-10:2020-11-20`\n\n### --output\n\nYou can specify the output file with `-o` parameter. For example `-o out.txt`\n\n### --rm\n\nYou can specify the `-rm` parameter to remove the downloaded archive folders after processing. This is useful to save disk space. \n\n### Demonstration Video\n\n[![Watch the video](https://i.imgur.com/J2CrvfM.png)](https://www.youtube.com/watch?v=Ct086YRm7i8)\n\n## The Speed Problem\n\nArchive.org throttles the speed when downloading files. Therefore, downloading an archive takes more time than usual. As a workaround, you can download the archives via Torrent and put them under the `archive/` folder which is located in the same directory with the urlhunter's binary. The directory tree will look like:\n\n```\n|-urlhunter\n|---urlhunter(binary)\n|---archive\n|-----urlteam_2020-11-20-11-17-04\n|-----urlteam_2020-11-17-11-17-04\n```\n\n## Example Use Cases\n\nurlhunter might be useful for cyber intelligence and bug bounty purposes. For example:\n\n`docs.google.com/a/acme.com` `drive.google.com/a/acme.com` keywords allow you to find public Google Docs\u0026Drive share links of Acme company.\n\n`acme.com,password_reset_token` keyword may allow you to find the working password reset tokens of acme.com\n\n`trello.com` allows you to find public Trello addresses.\n\n## Thanks\n\nSpecial thanks to Samet Bekmezci([@sametbekmezci](https://twitter.com/sametbekmezci)) who gave me the idea of this tool. \n","funding_links":[],"categories":["Go","[↑](#-table-of-contents) URLs","Weapons","Go (531)","bugbounty","\u003ca name=\"webdev\"\u003e\u003c/a\u003eWeb development","Pentesting","Other"],"sub_categories":["Tools","Reconnaissance"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Futkusen%2Furlhunter","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Futkusen%2Furlhunter","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Futkusen%2Furlhunter/lists"}