{"id":13910514,"url":"https://github.com/ynop/audiomate","last_synced_at":"2026-01-14T08:03:21.607Z","repository":{"id":27123146,"uuid":"112253012","full_name":"ynop/audiomate","owner":"ynop","description":"Python library for handling audio datasets.","archived":false,"fork":false,"pushed_at":"2023-07-06T21:54:14.000Z","size":9510,"stargazers_count":138,"open_issues_count":29,"forks_count":27,"subscribers_count":11,"default_branch":"master","last_synced_at":"2026-01-11T12:19:05.194Z","etag":null,"topics":["audio","audio-datasets","corpus-tools","data-loader","dataset-creation","dataset-filtering","dataset-manager","music","noise","speech","speech-recognition"],"latest_commit_sha":null,"homepage":"https://audiomate.readthedocs.io/","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":"mit","status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/ynop.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":"CONTRIBUTING.md","funding":null,"license":"LICENSE","code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null,"zenodo":null}},"created_at":"2017-11-27T21:54:21.000Z","updated_at":"2025-08-12T09:03:57.000Z","dependencies_parsed_at":"2024-01-08T12:17:48.432Z","dependency_job_id":null,"html_url":"https://github.com/ynop/audiomate","commit_stats":{"total_commits":487,"total_committers":12,"mean_commits":"40.583333333333336","dds":0.1416837782340863,"last_synced_commit":"080402eadaa81f77f64c8680510a2de64bc18e74"},"previous_names":[],"tags_count":10,"template":false,"template_full_name":null,"purl":"pkg:github/ynop/audiomate","repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/ynop%2Faudiomate","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/ynop%2Faudiomate/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/ynop%2Faudiomate/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/ynop%2Faudiomate/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/ynop","download_url":"https://codeload.github.com/ynop/audiomate/tar.gz/refs/heads/master","sbom_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/ynop%2Faudiomate/sbom","scorecard":null,"host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":286080680,"owners_count":28413527,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2026-01-14T05:26:33.345Z","status":"ssl_error","status_checked_at":"2026-01-14T05:21:57.251Z","response_time":107,"last_error":"SSL_connect returned=1 errno=0 peeraddr=140.82.121.5:443 state=error: unexpected eof while reading","robots_txt_status":"success","robots_txt_updated_at":"2025-07-24T06:49:26.215Z","robots_txt_url":"https://github.com/robots.txt","online":false,"can_crawl_api":true,"host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":["audio","audio-datasets","corpus-tools","data-loader","dataset-creation","dataset-filtering","dataset-manager","music","noise","speech","speech-recognition"],"created_at":"2024-08-07T00:01:30.728Z","updated_at":"2026-01-14T08:03:21.577Z","avatar_url":"https://github.com/ynop.png","language":"Python","readme":"# AUDIOMATE\n\n[![PyPI](https://img.shields.io/pypi/v/audiomate.svg)](https://pypi.python.org/pypi/audiomate)\n[![Build Status](https://travis-ci.com/ynop/audiomate.svg?branch=master)](https://travis-ci.com/ynop/audiomate)\n[![Documentation Status](https://readthedocs.org/projects/audiomate/badge/?version=latest)](https://audiomate.readthedocs.io/en/latest/?badge=latest)\n[![DeepSource](https://static.deepsource.io/deepsource-badge-light-mini.svg)](https://deepsource.io/gh/ynop/audiomate/?ref=repository-badge)\n[![DOI](https://joss.theoj.org/papers/10.21105/joss.02135/status.svg)](https://doi.org/10.21105/joss.02135)\n\nAudiomate is a library for easy access to audio datasets.\nIt provides the datastructures for accessing/loading different datasets in a generic way.\nThis should ease the use of audio datasets for example for machine learning tasks.\n\n```python\nimport audiomate\nfrom audiomate.corpus import io\n\n# Download a dataset\nesc_downloader = io.ESC50Downloader()\nesc_downloader.download('/local/path')\n\n# Load and work with the dataset\nesc50 = audiomate.Corpus.load('/local/path', reader='esc-50')\n\n# e.g. Read the audio signal and the label of specific sample/utterance\nutterance = esc50.utterances['1-100032-A-0']\nsamples = utterance.read_samples()\nlabel_list = utterance.label_lists[audiomate.corpus.LL_SOUND_CLASS]\n\nfor label in label_list:\n  print(label.start, label.value)\n```\n\nFurthermore it provides tools for interacting with datasets\n(validation, splitting, subsets, merge, filter), extracting features,\nfeeding samples for training ML models and more.\n\n* [Documentation](https://audiomate.readthedocs.io)\n* [Examples](https://github.com/ynop/audiomate/tree/master/examples)\n* [Changelog](https://audiomate.readthedocs.io/en/latest/notes/changelog.html)\n\nCurrently supported datasets:\n* [Acoustic Event Dataset](https://arxiv.org/pdf/1604.07160.pdf)\n* [AudioMNIST](https://github.com/soerenab/AudioMNIST)\n* [Mozilla Common Voice](https://voice.mozilla.org/)\n* [ESC-50](https://github.com/karoldvl/ESC-50)\n* [Fluent Speech Commands](http://www.fluent.ai/research/fluent-speech-commands/)\n* [Free Spoken Digit Dataset](https://github.com/Jakobovski/free-spoken-digit-dataset)\n* [German Distant Speech Corpus](https://www.inf.uni-hamburg.de/en/inst/ab/lt/resources/data/acoustic-models.html)\n* [Google Speech Commands](https://research.googleblog.com/2017/08/launching-speech-commands-dataset.html)\n* [GTZAN](http://marsyas.info/downloads/datasets.html)\n* [LibriSpeech](https://www.openslr.org/12/)\n* [M-AILABS Speech Dataset](https://www.caito.de/2019/01/the-m-ailabs-speech-dataset/)\n* [MUSAN](http://www.openslr.org/17/)\n* [LITIS Rouen Audio scene dataset](https://sites.google.com/site/alainrakotomamonjy/home/audio-scene)\n* [Spoken Wikipedia Corpora](https://nats.gitlab.io/swc/)\n* [Tatoeba](https://tatoeba.org/)\n* [TIMIT](https://github.com/philipperemy/timit)\n* [Urbansound8k](http://urbansounddataset.weebly.com/urbansound8k.html)\n* [Voxforge](http://www.voxforge.org/de)\n\nCurrently supported formats:\n* [Kaldi](http://kaldi-asr.org/)\n* [Mozilla DeepSpeech](https://github.com/mozilla/DeepSpeech)\n* [Wav2Letter](https://github.com/facebookresearch/wav2letter)\n* [NVIDIA Jasper](https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/SpeechRecognition/Jasper)\n* [Custom Formats](https://audiomate.readthedocs.io/en/latest/documentation/formats.html)\n\n## Installation\n\n```sh\npip install audiomate\n```\n\nInstall the latest development version:\n\n```sh\npip install git+https://github.com/ynop/audiomate.git\n```\n\n### Dependencies\n\n#### sox\nFor parts of the functionality (e.g. audio format conversion) [sox](http://sox.sourceforge.net) is used. In order to use it, you have to install sox.\n\n```sh\n# macos\nbrew install sox\n\n# with support for specific formats\nbrew install sox --with-lame --with-flac --with-libvorbis\n\n# linux\napt-get install sox\n\n# anaconda for macOS/windows/linux:\nconda install -c conda-forge sox\n```\n\n## Development\n\n### Prerequisites\n\n* [A supported version of Python \u003e 3.5](https://docs.python.org/devguide/index.html#status-of-python-branches)\n\nIt's recommended to use a virtual environment when developing audiomate.\nTo create one, execute the following command in the project's root directory:\n\n```\npython -m venv .\n```\n\nTo install audiomate and all it's dependencies, execute:\n\n```\npip install -e .\n```\n\n### Running the test suite\n\n```\npip install -e .[dev]\npytest\n```\n\nWith PyCharm you might have to change the default test runner. Otherwise, it might only suggest to use nose. To do so, go to File \u003e Settings \u003e Tools \u003e Python Integrated Tools (on the Mac it's PyCharm \u003e Preferences \u003e Settings \u003e Tools \u003e Python Integrated Tools) and change the test runner to py.test.\n\n### Benchmarks\n\nIn order to check the runtime of specific parts, ``pytest-benchmark`` is used. Benchmarks are normal test functions, but call the benchmark fixture for the code under test.\n\nTo run benchmarks:\n\n```\n# Run all\npytest bench\n\n# Specific benchmark\npytest bench/corpus/test_merge_corpus.py\n```\n\nTo compare between different runs:\n\n```\npytest-benchmark compare\n```\n\n### Editing the Documentation\n\nThe documentation is written in [reStructuredText](http://docutils.sourceforge.net/rst.html) and transformed into various output formats with the help of [Sphinx](http://www.sphinx-doc.org/).\n\n* [Syntax reference reStructuredText](http://docutils.sourceforge.net/docs/user/rst/quickref.html)\n* [Sphinx-specific additions to reStructuredText](http://www.sphinx-doc.org/en/stable/markup/index.html)\n\nTo generate the documentation, execute:\n\n```\npip install -e .[dev]\ncd docs\nmake html\n```\n\nThe generated files are written to `docs/_build/html`.\n\n### Versions\n\nVersions is handled using [bump2version](https://github.com/c4urself/bump2version). To bump the version:\n\n```\nbump2version [major,minor,patch,release,num]\n```\n\nIn order to directly go to a final relase version (skip .dev/.rc/...):\n\n```\nbump2version [major,minor,patch] --new-version x.x.x\n```\n\n### Release\n\nCommands to create a new release on pypi.\n\n```\nrm -rf build\nrm -rf dist\n\npython setup.py sdist\npython setup.py bdist_wheel\ntwine upload dist/*\n```\n","funding_links":[],"categories":["Python","Audio Related Packages"],"sub_categories":[],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fynop%2Faudiomate","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fynop%2Faudiomate","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fynop%2Faudiomate/lists"}