https://github.com/interactiveaudiolab/penn
Pitch Estimating Neural Networks (PENN)
https://github.com/interactiveaudiolab/penn
frequency music periodicity pitch speech voicing
Last synced: about 1 month ago
JSON representation
Pitch Estimating Neural Networks (PENN)
- Host: GitHub
- URL: https://github.com/interactiveaudiolab/penn
- Owner: interactiveaudiolab
- License: mit
- Created: 2021-01-10T03:50:25.000Z (almost 5 years ago)
- Default Branch: master
- Last Pushed: 2025-04-02T17:03:42.000Z (6 months ago)
- Last Synced: 2025-07-11T10:14:56.545Z (3 months ago)
- Topics: frequency, music, periodicity, pitch, speech, voicing
- Language: Python
- Homepage:
- Size: 73.3 MB
- Stars: 257
- Watchers: 10
- Forks: 24
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
Pitch-Estimating Neural Networks (PENN)
[](https://pypi.python.org/pypi/penn)
[](https://opensource.org/licenses/MIT)
[](https://pepy.tech/project/penn)Training, evaluation, and inference of neural pitch and periodicity estimators in PyTorch. Includes the original code for the paper ["Cross-domain Neural Pitch and Periodicity Estimation"](https://arxiv.org/abs/2301.12258).
## Table of contents
- [Installation](#installation)
- [Inference](#inference)
* [Application programming interface](#application-programming-interface)
* [`penn.from_audio`](#pennfrom_audio)
* [`penn.from_file`](#pennfrom_file)
* [`penn.from_file_to_file`](#pennfrom_file_to_file)
* [`penn.from_files_to_files`](#pennfrom_files_to_files)
* [Command-line interface](#command-line-interface)
- [Training](#training)
* [Download](#download)
* [Preprocess](#preprocess)
* [Partition](#partition)
* [Train](#train)
* [Monitor](#monitor)
- [Evaluation](#evaluation)
* [Evaluate](#evaluate)
* [Plot](#plot)
- [Citation](#citation)## Installation
If you want to perform pitch estimation using a pretrained FCNF0++ model, run
`pip install penn`If you want to train or use your own models, run
`pip install penn[train]`## Inference
Perform inference using FCNF0++
```
import penn# Load audio
audio, sample_rate = torchaudio.load('test/assets/gershwin.wav')# Here we'll use a 10 millisecond hopsize
hopsize = .01# Provide a sensible frequency range given your domain and model
fmin = 30.
fmax = 1000.# Choose a gpu index to use for inference. Set to None to use cpu.
gpu = 0# If you are using a gpu, pick a batch size that doesn't cause memory errors
# on your gpu
batch_size = 2048# Select a checkpoint to use for inference. Selecting None will
# download and use FCNF0++ pretrained on MDB-stem-synth and PTDB
checkpoint = None# Centers frames at hopsize / 2, 3 * hopsize / 2, 5 * hopsize / 2, ...
center = 'half-hop'# (Optional) Linearly interpolate unvoiced regions below periodicity threshold
interp_unvoiced_at = .065# (Optional) Select a decoding method. One of ['argmax', 'pyin', 'viterbi'].
decoder = 'viterbi'# Infer pitch and periodicity
pitch, periodicity = penn.from_audio(
audio,
sample_rate,
hopsize=hopsize,
fmin=fmin,
fmax=fmax,
checkpoint=checkpoint,
batch_size=batch_size,
center=center,
decoder=decoder,
interp_unvoiced_at=interp_unvoiced_at,
gpu=gpu)
```Note that pitch estimation is performed independently on each frame of audio. Then, a _decoding_ step occurs, which may or may not be computed independently on each frame. Most often, Viterbi decoding is used (as in, e.g., PYIN and CREPE). However, Viterbi decoding is slow. We made a fast Viterbi decoder called [torbi](https://github.com/maxrmorrison/torbi), which [we are working on adding to PyTorch](https://github.com/pytorch/pytorch/issues/121160). Until `torbi` is integrated into PyTorch (or otherwise made pip-installable), it is recommended to use the `dev` branch of `penn`, which uses `torbi` decoding by default, but is not pip-installable. Our paper [_Fine-Grained and Interpretable Neural Speech Editing_](https://www.maxrmorrison.com/sites/promonet/) introduces and demonstrates the efficacy of `torbi` for pitch decoding.
### Application programming interface
#### `penn.from_audio`
```
def from_audio(
audio: torch.Tensor,
sample_rate: int = penn.SAMPLE_RATE,
hopsize: float = penn.HOPSIZE_SECONDS,
fmin: float = penn.FMIN,
fmax: float = penn.FMAX,
checkpoint: Optional[Path] = None,
batch_size: Optional[int] = None,
center: str = 'half-window',
decoder: str = penn.DECODER,
interp_unvoiced_at: Optional[float] = None,
gpu: Optional[int] = None
) -> Tuple[torch.Tensor, torch.Tensor]:
"""Perform pitch and periodicity estimationArgs:
audio: The audio to extract pitch and periodicity from
sample_rate: The audio sample rate
hopsize: The hopsize in seconds
fmin: The minimum allowable frequency in Hz
fmax: The maximum allowable frequency in Hz
checkpoint: The checkpoint file
batch_size: The number of frames per batch
center: Padding options. One of ['half-window', 'half-hop', 'zero'].
interp_unvoiced_at: Specifies voicing threshold for interpolation
gpu: The index of the gpu to run inference onReturns:
pitch: torch.tensor(
shape=(1, int(samples // penn.seconds_to_sample(hopsize))))
periodicity: torch.tensor(
shape=(1, int(samples // penn.seconds_to_sample(hopsize))))
"""
```#### `penn.from_file`
```
def from_file(
file: Path,
hopsize: float = penn.HOPSIZE_SECONDS,
fmin: float = penn.FMIN,
fmax: float = penn.FMAX,
checkpoint: Optional[Path] = None,
batch_size: Optional[int] = None,
center: str = 'half-window',
decoder: str = penn.DECODER,
interp_unvoiced_at: Optional[float] = None,
gpu: Optional[int] = None
) -> Tuple[torch.Tensor, torch.Tensor]:
"""Perform pitch and periodicity estimation from audio on diskArgs:
file: The audio file
hopsize: The hopsize in seconds
fmin: The minimum allowable frequency in Hz
fmax: The maximum allowable frequency in Hz
checkpoint: The checkpoint file
batch_size: The number of frames per batch
center: Padding options. One of ['half-window', 'half-hop', 'zero'].
interp_unvoiced_at: Specifies voicing threshold for interpolation
gpu: The index of the gpu to run inference onReturns:
pitch: torch.tensor(shape=(1, int(samples // hopsize)))
periodicity: torch.tensor(shape=(1, int(samples // hopsize)))
"""
```#### `penn.from_file_to_file`
```
def from_file_to_file(
file: Path,
output_prefix: Optional[Path] = None,
hopsize: float = penn.HOPSIZE_SECONDS,
fmin: float = penn.FMIN,
fmax: float = penn.FMAX,
checkpoint: Optional[Path] = None,
batch_size: Optional[int] = None,
center: str = 'half-window',
decoder: str = penn.DECODER,
interp_unvoiced_at: Optional[float] = None,
gpu: Optional[int] = None
) -> None:
"""Perform pitch and periodicity estimation from audio on disk and saveArgs:
file: The audio file
output_prefix: The file to save pitch and periodicity without extension
hopsize: The hopsize in seconds
fmin: The minimum allowable frequency in Hz
fmax: The maximum allowable frequency in Hz
checkpoint: The checkpoint file
batch_size: The number of frames per batch
center: Padding options. One of ['half-window', 'half-hop', 'zero'].
interp_unvoiced_at: Specifies voicing threshold for interpolation
gpu: The index of the gpu to run inference on
"""
```#### `penn.from_files_to_files`
```
def from_files_to_files(
files: List[Path],
output_prefixes: Optional[List[Path]] = None,
hopsize: float = penn.HOPSIZE_SECONDS,
fmin: float = penn.FMIN,
fmax: float = penn.FMAX,
checkpoint: Optional[Path] = None,
batch_size: Optional[int] = None,
center: str = 'half-window',
decoder: str = penn.DECODER,
interp_unvoiced_at: Optional[float] = None,
num_workers: int = penn.NUM_WORKERS,
gpu: Optional[int] = None
) -> None:
"""Perform pitch and periodicity estimation from files on disk and saveArgs:
files: The audio files
output_prefixes: Files to save pitch and periodicity without extension
hopsize: The hopsize in seconds
fmin: The minimum allowable frequency in Hz
fmax: The maximum allowable frequency in Hz
checkpoint: The checkpoint file
batch_size: The number of frames per batch
center: Padding options. One of ['half-window', 'half-hop', 'zero'].
interp_unvoiced_at: Specifies voicing threshold for interpolation
num_workers: Number of CPU threads for async data I/O
gpu: The index of the gpu to run inference on
"""
```### Command-line interface
```
python -m penn
--files FILES [FILES ...]
[-h]
[--config CONFIG]
[--output_prefixes OUTPUT_PREFIXES [OUTPUT_PREFIXES ...]]
[--hopsize HOPSIZE]
[--fmin FMIN]
[--fmax FMAX]
[--checkpoint CHECKPOINT]
[--batch_size BATCH_SIZE]
[--center {half-window,half-hop,zero}]
[--decoder {argmax,pyin,viterbi}]
[--interp_unvoiced_at INTERP_UNVOICED_AT]
[--num_workers NUM_WORKERS]
[--gpu GPU]required arguments:
--files FILES [FILES ...]
The audio files to processoptional arguments:
-h, --help
show this help message and exit
--config CONFIG
The configuration file. Defaults to using FCNF0++.
--output_prefixes OUTPUT_PREFIXES [OUTPUT_PREFIXES ...]
The files to save pitch and periodicity without extension.
Defaults to files without extensions.
--hopsize HOPSIZE
The hopsize in seconds. Defaults to 0.01 seconds.
--fmin FMIN
The minimum frequency allowed in Hz. Defaults to 31.0 Hz.
--fmax FMAX
The maximum frequency allowed in Hz. Defaults to 1984.0 Hz.
--checkpoint CHECKPOINT
The model checkpoint file. Defaults to ./penn/assets/checkpoints/fcnf0++.pt.
--batch_size BATCH_SIZE
The number of frames per batch. Defaults to 2048.
--center {half-window,half-hop,zero}
Padding options
--decoder {argmax,pyin,viterbi}
Posteriorgram decoder
--interp_unvoiced_at INTERP_UNVOICED_AT
Specifies voicing threshold for interpolation. Defaults to 0.1625.
--num_workers
Number of CPU threads for async data I/O
--gpu GPU
The index of the gpu to perform inference on. Defaults to CPU.
```## Training
### Download
`python -m penn.data.download`
Downloads and uncompresses the `mdb` and `ptdb` datasets used for training.
### Preprocess
`python -m penn.data.preprocess --config `
Converts each dataset to a common format on disk ready for training. You
can optionally pass a configuration file to override the default configuration.### Partition
`python -m penn.partition`
Generates `train`, `valid`, and `test` partitions for `mdb` and `ptdb`.
Partitioning is deterministic given the same random seed. You do not need to
run this step, as the original partitions are saved in
`penn/assets/partitions`.### Train
`python -m penn.train --config --gpu `
Trains a model according to a given configuration on the `mdb` and `ptdb`
datasets.### Monitor
You can monitor training via `tensorboard`.
```
tensorboard --logdir runs/ --port --load_fast true
```To use the `torchutil` notification system to receive notifications for long
jobs (download, preprocess, train, and evaluate), set the
`PYTORCH_NOTIFICATION_URL` environment variable to a supported webhook as
explained in [the Apprise documentation](https://pypi.org/project/apprise/).## Evaluation
### Evaluate
```
python -m penn.evaluate \
--config \
--checkpoint \
--gpu
```Evaluate a model. `` is the checkpoint file to evaluate and ``
is the GPU index.### Plot
```
python -m penn.plot.density \
--config \
--true_datasets \
--inference_datasets \
--output_file \
--checkpoint \
--gpu
```Plot the data distribution and inferred distribution for a given dataset and
save to a jpg file.```
python -m penn.plot.logits \
--config \
--audio_file \
--output_file \
--checkpoint \
--gpu
```Plot the pitch posteriorgram of an audio file and save to a jpg file.
```
python -m penn.plot.threshold \
--names \
--evaluations \
--output_file
```Plot the periodicity performance (voiced/unvoiced F1) over mdb and ptdb as a
function of the voiced/unvoiced threshold. `names` are the plot labels to give
each evaluation. `evaluations` are the names of the evaluations to plot.## Citation
### IEEE
M. Morrison, C. Hsieh, N. Pruyne, and B. Pardo, "Cross-domain Neural Pitch and Periodicity Estimation," arXiv preprint arXiv:2301.12258, 2023.### BibTex
```
@inproceedings{morrison2023cross,
title={Cross-domain Neural Pitch and Periodicity Estimation},
author={Morrison, Max and Hsieh, Caedon and Pruyne, Nathan and Pardo, Bryan},
booktitle={arXiv preprint arXiv:2301.12258},
year={2023}
}