Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/newrelic/rusty-hog

A suite of secret scanners built in Rust for performance. Based on TruffleHog (https://github.com/dxa4481/truffleHog) which is written in Python.
https://github.com/newrelic/rusty-hog

Last synced: about 1 month ago
JSON representation

A suite of secret scanners built in Rust for performance. Based on TruffleHog (https://github.com/dxa4481/truffleHog) which is written in Python.

Awesome Lists containing this project

README

        

Rusty Hog is a secret scanner built in Rust for performance, and based on TruffleHog which is written
in Python. Rusty Hog provides the following binaries:

* Ankamali Hog: Scans for secrets in a Google doc.
* Berkshire Hog: Scans for secrets in an S3 bucket.
* Choctaw Hog: Scans for secrets in a Git repository.
* Duroc Hog: Scans for secrets in a directory, file, and archive.
* Essex Hog: Scans for secrets in a Confluence wiki page.
* Gottingen Hog: Scans for secrets in a JIRA issue.
* Slack Hog: Scans for secrets in a Slack Channel.

## Table of contents

- [Usage](#usage)
- [How to install using downloaded binaries](#how-to-install-using-downloaded-binaries)
- [How to run using DockerHub](#how-to-run-using-dockerhub)
- [How to build](#how-to-build)
- [How to build on Windows](#how-to-build-on-windows)
- [Anakamali Hog (GDoc Scanner) usage](#anakamali-hog-gdoc-scanner-usage)
- [Berkshire Hog (S3 Scanner - CLI) usage](#berkshire-hog-s3-scanner---cli-usage)
- [Berkshire Hog (S3 Scanner - Lambda) usage](#berkshire-hog-s3-scanner---lambda-usage)
- [Choctaw Hog (Git Scanner) usage](#choctaw-hog-git-scanner-usage)
- [Duroc Hog (file system scanner) usage](#duroc-hog-file-system-scanner-usage)
- [Essex Hog (Confluence scanner) usage](#essex-hog-confluence-scanner-usage)
- [Gottingen Hog (JIRA scanner) usage](#gottingen-hog-jira-scanner-usage)
- [Hante Hog (SLACK scanner) usage](#slack-hog-slack-scanner-usage)
- [Regex JSON file format](#regex-json-file-format)
- [Allowlist JSON file format](#allowlist-json-file-format)
- [Project information](#project-information)
- [Open source license](#open-source-license)
- [Support](#support)
- [Community](#community)
- [Issues / enhancement requests](#issues--enhancement-requests)
- [Contributing](#contributing)
- [Feature Roadmap](#feature-roadmap)
- [What does the name mean?](#what-does-the-name-mean)

# Usage

This project provides a set of scanners that use regular expressions to try and detect the presence of sensitive
information, such as API keys, passwords, and personal information. It includes a set of regular expressions by
default, but also accepts a JSON object containing your custom regular expressions.

## How to install using downloaded binaries
Download and unzip the [latest ZIP](https://github.com/newrelic/rusty-hog/releases/)
on the releases tab. Then, run each binary with `-h` to see the usage.

```shell script
wget https://github.com/newrelic/rusty-hog/releases/download/v1.0.11/rustyhogs-darwin-choctaw_hog-1.0.11.zip
unzip rustyhogs-darwin-choctaw_hog-1.0.11.zip
darwin_releases/choctaw_hog -h
```

## How to run using DockerHub
Rusty Hog Docker images can be found at the authors personal DockerHub page [here](https://hub.docker.com/u/wetfeet2000)
A Docker Image is built for each Hog and for each release. So to use choctaw_hog you would run the following commands:

```shell script
docker pull wetfeet2000/choctaw_hog:1.0.10
docker run -it --rm wetfeet2000/choctaw_hog:1.0.10 --help
```

## How to build
- Ensure you have [Rust](https://www.rust-lang.org/learn/get-started) installed and on your path.
- Clone this repo, and then run `cargo build --release`. The binaries are located in `target/release`.
- To build and view HTML documents, run ```cargo doc --no-deps --open```.
- To run unit tests, run ```cargo test```.
- To cross-compile Berkshire Hog for the AWS Lambda environment, run the following commands and upload berkshire_lambda.zip to
your AWS Lambda dashboard:
```shell script
docker run --rm -it -v "$(pwd)":/home/rust/src ekidd/rust-musl-builder cargo build --release
cp target/x86_64-unknown-linux-musl/release/berkshire_hog bootstrap
zip -j berkshire_lambda.zip bootstrap
```

## How to build on Windows
You will need to compile static OpenSSL binaries and tell Rust/Cargo where to find them:
```
mkdir \Tools
cd \Tools
git clone https://github.com/Microsoft/vcpkg.git
cd vcpkg
.\bootstrap-vcpkg.bat
.\vcpkg.exe install openssl:x64-windows-static

$env:OPENSSL_DIR = 'C:\Tools\vcpkg\installed\x64-windows-static'
$env:OPENSSL_STATIC = 'Yes'
[System.Environment]::SetEnvironmentVariable('OPENSSL_DIR', $env:OPENSSL_DIR, [System.EnvironmentVariableTarget]::User)
[System.Environment]::SetEnvironmentVariable('OPENSSL_STATIC', $env:OPENSSL_STATIC, [System.EnvironmentVariableTarget]::User)
```
You can now follow the main build instructions listed above.

## How to build Berkshire Hog for Lambda

### macOS

Use [Homebrew](https://brew.sh) to get the dependencies:

```
brew install rpm2cpio FiloSottile/musl-cross/musl-cross
```

Then run `./build_lambda_macos.sh`.

The build script will build against OpenSSL 3.0.12. Use `export OPENSSL_BUILD_VER=3.0.12` to override.

The build script will build against Amazon Linux kernel headers
provided by their RPM; `export AMAZON_KERNEL_HEADERS_RPM_URL=...` to override where the RPM is downloaded from. (There's nothing preventing a different distribution's linux-headers RPM from being used, we just need the linux-headers to build openssl for Linux)

The build script will create a build-deps directory in your current source root. You can `rm -rf` this directory safely, but it will be recreated on the next build script run. It will also do various consistency checks to make sure the build should work and if those fail, might ask you to `rm -rf` that directory anyway to try again.

### Linux

Make sure `cross` is installed (`cargo install cross`), then just `./build_lambda.sh`.

# Commands

## Anakamali Hog (GDoc Scanner) usage
```
USAGE:
ankamali_hog [FLAGS] [OPTIONS]

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--oauthsecret Path to an OAuth secret file (JSON) ./clientsecret.json by default
--oauthtoken Path to an OAuth token storage file ./temp_token by default
--prettyprint Outputs the JSON in human readable format
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
-a, --allowlist Sets a custom allowlist JSON file
--default_entropy_threshold Default entropy threshold (0.6 by default)
-o, --outputfile Sets the path to write the scanner results to (stdout by default)

--regex Sets a custom regex JSON file

ARGS:
The ID of the Google drive file you want to scan
```

## Berkshire Hog (S3 Scanner - CLI) usage
```
USAGE:
berkshire_hog [FLAGS] [OPTIONS]

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--prettyprint Outputs the JSON in human readable format
-r, --recursive Recursively scans files under the prefix
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
-a, --allowlist Sets a custom allowlist JSON file
--default_entropy_threshold Default entropy threshold (0.6 by default)
-o, --outputfile Sets the path to write the scanner results to (stdout by default)

--profile When using a configuration file, enables a non-default profile

--regex Sets a custom regex JSON file

ARGS:
The location of a S3 bucket and optional prefix or filename to scan. This must be written in the
form s3://mybucket[/prefix_or_file]
Sets the region of the S3 bucket to scan
```

## Berkshire Hog (S3 Scanner - Lambda) usage
Berkshire Hog is currently designed to be used as a Lambda function. This is the basic data flow:


┌───────────┐ ┌───────┐ ┌────────────────┐ ┌────────────┐
│ S3 bucket │ ┌────────┐ │ │ │ Berkshire Hog │ │ S3 bucket │
│ (input) ─┼─┤S3 event├──▶│ SQS │────▶│ (Lambda) │────▶│ (output) │
│ │ └────────┘ │ │ │ │ │ │
└───────────┘ └───────┘ └────────────────┘ └────────────┘

In order to run Berkshire Hog this way, set up the following:
1) Configure the input bucket to send an "event" to SQS for each PUSH/PUT event.
2) Set up the SQS topic to accept events from S3, including IAM permissions.
3) Run Berkshire Hog with IAM access to SQS and S3.

## Choctaw Hog (Git Scanner) usage
```
USAGE:
choctaw_hog [FLAGS] [OPTIONS]

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--prettyprint Outputs the JSON in human readable format
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
--default_entropy_threshold Default entropy threshold (4.5 by default)
--httpspass Takes a password for HTTPS-based authentication
--httpsuser Takes a username for HTTPS-based authentication
-o, --outputfile Sets the path to write the scanner results to (stdout by default)
--recent_days Filters commits to the last number of days (branch agnostic)
-r, --regex Sets a custom regex JSON file
--since_commit Filters commits based on date committed (branch agnostic)
--sshkeypath Takes a path to a private SSH key for git authentication, defaults to ssh-agent
--sshkeyphrase Takes a passphrase to a private SSH key for git authentication, defaults to none
--until_commit Filters commits based on date committed (branch agnostic)
-a, --allowlist Sets a custom ALLOWLIST JSON file

ARGS:
Sets the path (or URL) of the Git repo to scan. SSH links must include username (git@)
```

## Duroc Hog (file system scanner) usage
```
USAGE:
duroc_hog [FLAGS] [OPTIONS]

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--norecursive Disable recursive scanning of all subdirectories underneath the supplied path
--prettyprint Outputs the JSON in human readable format
-z, --unzip Recursively scans archives (ZIP and TAR) in memory (dangerous)
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
-a, --allowlist Sets a custom allowlist JSON file
--default_entropy_threshold Default entropy threshold (0.6 by default)
-o, --outputfile Sets the path to write the scanner results to (stdout by default)
-r, --regex Sets a custom regex JSON file

ARGS:
Sets the path of the directory or file to scan.
```

## Essex Hog (Confluence scanner) usage
```
USAGE:
essex_hog [FLAGS] [OPTIONS]

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--prettyprint Outputs the JSON in human readable format
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
-a, --allowlist Sets a custom allowlist JSON file
--authtoken Confluence basic auth bearer token (instead of user & pass)

--default_entropy_threshold Default entropy threshold (0.6 by default)
-o, --outputfile Sets the path to write the scanner results to (stdout by default)
--password Confluence password (crafts basic auth header)
--regex Sets a custom regex JSON file
--username Confluence username (crafts basic auth header)

ARGS:
The ID (e.g. 1234) of the confluence page you want to scan
Base URL of Confluence instance (e.g. https://newrelic.atlassian.net/)
```

## Gottingen Hog (JIRA scanner) usage
```
Jira secret scanner in Rust.

USAGE:
gottingen_hog [FLAGS] [OPTIONS]

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--prettyprint Outputs the JSON in human readable format
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
-a, --allowlist Sets a custom allowlist JSON file
--authtoken Jira basic auth bearer token (instead of user & pass)
--default_entropy_threshold Default entropy threshold (0.6 by default)
--url Base URL of JIRA instance (e.g. https://jira.atlassian.net/)
-o, --outputfile Sets the path to write the scanner results to (stdout by default)
--password Jira password (crafts basic auth header)
--regex Sets a custom regex JSON file
--username Jira username (crafts basic auth header)

ARGS:
The ID (e.g. PROJECT-123) of the Jira issue you want to scan
```

## Hante Hog (SLACK scanner) usage
```
Slack secret scanner in Rust.

USAGE:
hante_hog [FLAGS] [OPTIONS] --authtoken --channelid --url

FLAGS:
--caseinsensitive Sets the case insensitive flag for all regexes
--entropy Enables entropy scanning
--prettyprint Outputs the JSON in human readable format
-v, --verbose Sets the level of debugging information
-h, --help Prints help information
-V, --version Prints version information

OPTIONS:
-a, --allowlist Sets a custom allowlist JSON file
--authtoken Slack basic auth bearer token
--channelid
The ID (e.g. C12345) of the Slack channel you want to scan

--default_entropy_threshold Default entropy threshold (0.6 by default)
--latest End of time range of messages to include in search
--oldest Start of time range of messages to include in search
-o, --outputfile
Sets the path to write the scanner results to (stdout by default)

--regex Sets a custom regex JSON file
--url
Base URL of Slack Workspace (e.g. https://[WORKSPACE NAME].slack.com)
```

## Regex JSON file format

The regex option on scanners allows users to provide a path to their own JSON file of regular
expressions that match sensitive material. Any provided file currently will replace, not append to, the default
regular expressions provided by SecretScanner. The expected format of the file is a single json object.

The keys should be names for the type of secret each regex entry will detect, as the keys will be used for the reason
properties output by the scanner.

Each value should be a string containing a valid [https://docs.rs/regex/1.3.9/regex/#syntax](regular expression for Rust)
that should match the type of secret described by its corresponding key.

As of version 1.0.8, the Rusty Hog engine also supports objects as values for each secret.
The object can contain all of the following:

- a pattern property with the matching regex expression (mandatory)
- an entropy_filter property with a boolean value to enable entropy scanning for this information (mandatory)
- a threshold property to customize the entropy tolerance on a scale of 0 - 1 (optional, will adjust for old 1-8 format, default 0.6)
- a keyspace property to indicate how many possible values are in the key, e.g. 16 for hex, 64 for base64, 128 for ASCII (optional, default 128)
- a make_ascii_lowercase property to indicate whether Rust should perform .make_ascii_lowercase() on the key before calculating entropy (optional, default false)

The higher the threshold, the more entropy is required in the secret to consider it a match.

An example of this format is here:

```json
{
"Generic Secret": {
"pattern": "(?i)secret[\\s[[:punct:]]]{1,4}[0-9a-zA-Z-_]{16,64}[\\s[[:punct:]]]?",
"entropy_filter": true,
"threshold": "0.6"
},
"Slack Token": {
"pattern": "(xox[p|b|o|a]-[0-9]{12}-[0-9]{12}-[0-9]{12}-[a-z0-9]{32})",
"entropy_filter": true,
"threshold": "0.6",
"keyspace": "36",
"make_ascii_lowercase": true
},
"Google API Key": {
"pattern": "AIza[0-9A-Za-z\\-_]{35}",
"entropy_filter": true
},
"PGP private key block": "-----BEGIN PGP PRIVATE KEY BLOCK-----"
}
```

As of version 1.0.11, the current default regex JSON used is as follows:

```json
{
"Slack Token": "(xox[p|b|o|a]-[0-9]{12}-[0-9]{12}-[0-9]{12}-[a-z0-9]{32})",
"RSA private key": "-----BEGIN RSA PRIVATE KEY-----",
"SSH (DSA) private key": "-----BEGIN DSA PRIVATE KEY-----",
"SSH (EC) private key": "-----BEGIN EC PRIVATE KEY-----",
"PGP private key block": "-----BEGIN PGP PRIVATE KEY BLOCK-----",
"Amazon AWS Access Key ID": "AKIA[0-9A-Z]{16}",
"Amazon MWS Auth Token": "amzn\\.mws\\.[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}",
"Facebook Access Token": "EAACEdEose0cBA[0-9A-Za-z]+",
"Facebook OAuth": "(?i)facebook[\\s[[:punct:]]]{1,4}[0-9a-f]{32}[\\s[[:punct:]]]?",
"GitHub": "(?i)(github|access[[:punct:]]token)[\\s[[:punct:]]]{1,4}[0-9a-zA-Z]{35,40}",
"Generic API Key": {
"pattern": "(?i)(api|access)[\\s[[:punct:]]]?key[\\s[[:punct:]]]{1,4}[0-9a-zA-Z\\-_]{16,64}[\\s[[:punct:]]]?",
"entropy_filter": true,
"threshold": "0.6",
"keyspace": "guess"
},
"Generic Account API Key": {
"pattern": "(?i)account[\\s[[:punct:]]]?api[\\s[[:punct:]]]{1,4}[0-9a-zA-Z\\-_]{16,64}[\\s[[:punct:]]]?",
"entropy_filter": true,
"threshold": "0.6",
"keyspace": "guess"
},
"Generic Secret": {
"pattern": "(?i)secret[\\s[[:punct:]]]{1,4}[0-9a-zA-Z-_]{16,64}[\\s[[:punct:]]]?",
"entropy_filter": true,
"threshold": "0.6",
"keyspace": "guess"
},
"Google API Key": "AIza[0-9A-Za-z\\-_]{35}",
"Google Cloud Platform API Key": "AIza[0-9A-Za-z\\-_]{35}",
"Google Cloud Platform OAuth": "(?i)[0-9]+-[0-9A-Za-z_]{32}\\.apps\\.googleusercontent\\.com",
"Google Drive API Key": "AIza[0-9A-Za-z\\-_]{35}",
"Google Drive OAuth": "(?i)[0-9]+-[0-9A-Za-z_]{32}\\.apps\\.googleusercontent\\.com",
"Google (GCP) Service-account": "(?i)\"type\": \"service_account\"",
"Google Gmail API Key": "AIza[0-9A-Za-z\\-_]{35}",
"Google Gmail OAuth": "(?i)[0-9]+-[0-9A-Za-z_]{32}\\.apps\\.googleusercontent\\.com",
"Google OAuth Access Token": "ya29\\.[0-9A-Za-z\\-_]+",
"Google YouTube API Key": "AIza[0-9A-Za-z\\-_]{35}",
"Google YouTube OAuth": "(?i)[0-9]+-[0-9A-Za-z_]{32}\\.apps\\.googleusercontent\\.com",
"Heroku API Key": "[h|H][e|E][r|R][o|O][k|K][u|U][\\s[[:punct:]]]{1,4}[0-9A-F]{8}-[0-9A-F]{4}-[0-9A-F]{4}-[0-9A-F]{4}-[0-9A-F]{12}",
"MailChimp API Key": "[0-9a-f]{32}-us[0-9]{1,2}",
"Mailgun API Key": "(?i)key-[0-9a-zA-Z]{32}",
"Credentials in absolute URL": "(?i)((https?|ftp)://)(([a-z0-9$_\\.\\+!\\*'\\(\\),;\\?&=-]|%[0-9a-f]{2})+(:([a-z0-9$_\\.\\+!\\*'\\(\\),;\\?&=-]|%[0-9a-f]{2})+)@)((([a-z0-9]\\.|[a-z0-9][a-z0-9-]*[a-z0-9]\\.)*[a-z][a-z0-9-]*[a-z0-9]|((\\d|[1-9]\\d|1\\d{2}|2[0-4][0-9]|25[0-5])\\.){3}(\\d|[1-9]\\d|1\\d{2}|2[0-4][0-9]|25[0-5]))(:\\d+)?)(((/+([a-z0-9$_\\.\\+!\\*'\\(\\),;:@&=-]|%[0-9a-f]{2})*)*(\\?([a-z0-9$_\\.\\+!\\*'\\(\\),;:@&=-]|%[0-9a-f]{2})*)?)?)?",
"PayPal Braintree Access Token": "(?i)access_token\\$production\\$[0-9a-z]{16}\\$[0-9a-f]{32}",
"Picatic API Key": "(?i)sk_live_[0-9a-z]{32}",
"Slack Webhook": "(?i)https://hooks.slack.com/services/T[a-zA-Z0-9_]{8}/B[a-zA-Z0-9_]{8}/[a-zA-Z0-9_]{24}",
"Stripe API Key": "(?i)sk_live_[0-9a-zA-Z]{24}",
"Stripe Restricted API Key": "(?i)rk_live_[0-9a-zA-Z]{24}",
"Square Access Token": "(?i)sq0atp-[0-9A-Za-z\\-_]{22}",
"Square OAuth Secret": "(?i)sq0csp-[0-9A-Za-z\\-_]{43}",
"Twilio API Key": "SK[0-9a-fA-F]{32}",
"Twitter Access Token": "(?i)twitter[\\s[[:punct:]]]{1,4}[1-9][0-9]+-[0-9a-zA-Z]{40}",
"Twitter OAuth": "(?i)twitter[\\s[[:punct:]]]{1,4}['|\"]?[0-9a-zA-Z]{35,44}['|\"]?",
"New Relic Partner & REST API Key": "[\\s[[:punct:]]][A-Fa-f0-9]{47}[\\s[[:punct:]][[:cntrl:]]]",
"New Relic Mobile Application Token": "[\\s[[:punct:]]][A-Fa-f0-9]{42}[\\s[[:punct:]][[:cntrl:]]]",
"New Relic Synthetics Private Location": "(?i)minion_private_location_key",
"New Relic Insights Key (specific)": "(?i)insights[\\s[[:punct:]]]?(key|query|insert)[\\s[[:punct:]]]{1,4}\\b[\\w-]{32,40}\\b",
"New Relic Insights Key (vague)": "(?i)(query|insert)[\\s[[:punct:]]]?key[\\s[[:punct:]]]{1,4}b[\\w-]{32,40}\\b",
"New Relic License Key": "(?i)license[\\s[[:punct:]]]?key[\\s[[:punct:]]]{1,4}\\b[\\w-]{32,40}\\b",
"New Relic Internal API Key": "(?i)nr-internal-api-key",
"New Relic HTTP Auth Headers and API Key": "(?i)(x|newrelic|nr)-?(admin|partner|account|query|insert|api|license)-?(id|key)[\\s[[:punct:]]]{1,4}\\b[\\w-]{32,47}\\b",
"New Relic API Key Service Key (new format)": "(?i)NRAK-[A-Z0-9]{27}",
"New Relic APM License Key (new format)": "(?i)[a-f0-9]{36}NRAL",
"New Relic APM License Key (new format, region-aware)": "(?i)[a-z]{2}[0-9]{2}xx[a-f0-9]{30}NRAL",
"New Relic REST API Key (new format)": "(?i)NRRA-[a-f0-9]{42}",
"New Relic Admin API Key (new format)": "(?i)NRAA-[a-f0-9]{27}",
"New Relic Insights Insert Key (new format)": "(?i)NRII-[A-Za-z0-9-_]{32}",
"New Relic Insights Query Key (new format)": "(?i)NRIQ-[A-Za-z0-9-_]{32}",
"New Relic Synthetics Private Location Key (new format)": "(?i)NRSP-[a-z]{2}[0-9]{2}[a-f0-9]{31}",
"Email address": "(?i)\\b(?:[a-z0-9!#$%&'*+/=?^_`{|}~-]+(?:\\.[a-z0-9!#$%&'*+/=?^_`{|}~-]+)*)@[a-z0-9][a-z0-9-]+\\.(com|de|cn|net|uk|org|info|nl|eu|ru)([\\W&&[^:/]]|\\A|\\z)",
"New Relic Account IDs in URL": "(newrelic\\.com/)?accounts/\\d{1,10}/",
"Account ID": "(?i)account[\\s[[:punct:]]]?id[\\s[[:punct:]]]{1,4}\\b[\\d]{1,10}\\b",
"Salary Information": "(?i)(salary|commission|compensation|pay)([\\s[[:punct:]]](amount|target))?[\\s[[:punct:]]]{1,4}\\d+"
}
```

## Allowlist JSON file format

Scanners provide an allowlist feature. This allows you to specify a list of regular expressions for each pattern that
will be ignored by the scanner. You can now optionally supply a list of regular expressions that are evaluated against
the file path as well.

The format for this allowlist file should be a single json object. Each key in the allowlist should match a key in the
regex json, and the value can be one of two things:
1) An array of strings that are exceptions for that regex pattern. For example:
2) An object with at least one key (patterns) and optionally a second key (paths).

In addition, you can specify the key `` which is evaluated against all patterns.

The following is the default allowlist included in all scans:

```json
{
"Email address": {
"patterns": [
"(?i)@newrelic.com",
"(?i)noreply@",
"(?i)test@"
],
"paths": [
"(?i)authors",
"(?i)contributors",
"(?i)license",
"(?i)maintainers",
"(?i)third_party_notices"
]
},
"Credentials in absolute URL": {
"patterns": [
"(?i)(https?://)?user:pass(word)?@"
]
},
"New Relic API Key Service Key (new format)": {
"patterns": [
"NRAK-123456789ABCDEFGHIJKLMNOPQR"
]
},
"Generic API Key": {
"patterns": [
"(?i)sanitizeAPIKeyForLogging"
]
},
"New Relic License Key": {
"patterns": [
"(?i)bootstrap_newrelic_admin_license_key",
"(?i)xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
"(?i)__YOUR_NEW_RELIC_LICENSE_KEY__LICENSE__",
"(?i)YOUR_NEW_RELIC_APPLICATION_TOKEN"
]
},
"Generic Secret": {
"patterns": [
"(?i)secret:NewRelicLicenseKeySecret"
]
},
"": [
"(?i)example",
"(?i)fake",
"(?i)replace",
"(?i)deadbeef",
"(?i)ABCDEFGHIJKLMNOPQRSTUVWX",
"1234567890"
]
}
```

Be aware that in these are strings, not regex expressions, and the keys for this allowlist have to a key in the regex json.
Keys are case-sensitive.

# Project information
## Open source license

This project is distributed under the [Apache 2 license](LICENSE).

## Support

New Relic has open-sourced this project. This project is provided AS-IS WITHOUT WARRANTY OR SUPPORT, although you can report issues and contribute to the project here on GitHub.

_Please do not report issues with this software to New Relic Global Technical Support._

## Community

New Relic hosts and moderates an online forum where customers can interact with New Relic employees as well as other customers to get help and share best practices. Like all official New Relic open source projects, there's a related Community topic in the New Relic Explorer's Hub. You can find this project's topic/threads here:

https://discuss.newrelic.com/t/rusty-hog-multi-platform-secret-key-scanner/90117

## Issues / enhancement requests

Submit issues and enhancement requests in the [Issues tab of this repository](../../issues). Please search for and review the existing open issues before submitting a new issue.

## Contributing

Contributions are welcome (and if you submit a enhancement request, expect to be invited to contribute it yourself). Please review our [Contributors Guide](CONTRIBUTING.md).

Keep in mind that when you submit your pull request, you'll need to sign the CLA via the click-through using CLA-Assistant. If you'd like to execute our corporate CLA, or if you have any questions, please drop us an email at [email protected].

## Feature Roadmap

- 1.1: Enterprise features
- [ ] Support config files (instead of command line args)
- [ ] Support environment variables instead of CLI args
- [ ] Multi-threading
- [ ] Better context detection and false positive filtering (GitHound, machine learning)
- [ ] Use Rusoto instead of s3-rust
- [x] Add JIRA scanner
- [x] Add file-system & archive scanner
- [ ] Use Rust features to reduce compilation dependencies?

- 1.2: Integration with larger scripts and UIs
- [ ] Support Github API for larger org management
- [ ] Scan all repos for a list of users
- [x] Scan all repos in an org
- [ ] Generate a web report or web interface. Support "save state" generation from UI.
- [ ] Agent/manager model
- [ ] Scheduler process (blocked by save state support)

## What does the name mean?
TruffleHog is considered the de facto standard / original secret scanner. I have been
building a suite of secret scanning tools for various platforms based on TruffleHog
and needed a naming scheme, so I started at the top of Wikipedia's
[list of pig breeds](https://en.wikipedia.org/wiki/List_of_pig_breeds).
Thus each tool name is a breed of pig starting at "A" and working up.