Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/audiolabs/blind-multi-room-model
Repository accompanying our paper: "A Multi-Room Transition Dataset for Blind Estimation of Energy Decay" (IWAENC'2024)
https://github.com/audiolabs/blind-multi-room-model
Last synced: 9 days ago
JSON representation
Repository accompanying our paper: "A Multi-Room Transition Dataset for Blind Estimation of Energy Decay" (IWAENC'2024)
- Host: GitHub
- URL: https://github.com/audiolabs/blind-multi-room-model
- Owner: audiolabs
- License: mit
- Created: 2024-08-09T08:12:30.000Z (3 months ago)
- Default Branch: main
- Last Pushed: 2024-08-09T08:54:54.000Z (3 months ago)
- Last Synced: 2024-08-09T10:15:55.865Z (3 months ago)
- Language: Python
- Size: 16.5 MB
- Stars: 0
- Watchers: 4
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# A Multi-Room Transition Dataset for Blind Estimation of Energy Decay
Philipp Götz1, Georg Götz2, Nils Meyer-Kahlen2, Kyung Yun Lee2, Karolina Prawda2, Emanuël A. P. Habets1, and Sebastian J. Schlecht3
_1International Audio Laboratories Erlangen, Germany\
2Acoustics Lab, Dpt. of Information and Communications Engineering, Aalto University, Finland\
3Friedrich-Alexander-University Erlangen-Nuremberg (FAU), Germany_---
The multimodal dataset described in the IWAENC 2024 publication, including room impulse responses (RIRs) and 360˚-photos of each measurement position, is hosted at Zenodo (https://zenodo.org/records/11388246).This is the accompanying code repository for the blind energy decay function (EDF) estimation method proposed in the paper and includes the source code of the model, the training routines and some basic visualizing of the results.
## Installation
[![python](https://img.shields.io/badge/-Python_3.10-blue?logo=python&logoColor=white)](https://github.com/pre-commit/pre-commit)
[![pytorch](https://img.shields.io/badge/PyTorch_2.0+-ee4c2c?logo=pytorch&logoColor=white)](https://pytorch.org/get-started/locally/)
[![lightning](https://img.shields.io/badge/-Lightning_2.0+-792ee5?logo=pytorchlightning&logoColor=white)](https://pytorchlightning.ai/)
[![hydra](https://img.shields.io/badge/Config-Hydra_1.3-89b8cd)](https://hydra.cc/)
[![license](https://img.shields.io/badge/License-MIT-green.svg?labelColor=gray)](https://github.com/ashleve/lightning-hydra-template#license)The Python packages required to run the code can be installed using the requirements.txt file
```
pip install -f requirements.txt
```
The model training and evaluation data is hosted at a Google-Drive and downloaded by running ``download.sh``. Depending on wether a GPU is available or not, a softlink to ``cpu.yaml`` or ``gpu.yaml`` can be created in ``configs/local``.
## Training
The pre-generated dataset from the Google-Drive is constructed as described in the paper. In a preliminary step the blind T60 estimator is trained by running
```
python src/train.py -cn train model=baseline_t60 hydra=baseline_t60
```
Upon convergence, the trained model is used as a baseline which computes linear EDCs from blind T60 estimates. As an additional, non-blind baseline (not based on speech but on RIRs directly), a pre-trained _DecayFitNet_ is used to generate multi-slope EDCs. The blind EDC estimator is trained using
```
python src/train.py -cn train model=baseline_edc
```
More information on DecayFitNet can be found at https://github.com/georg-goetz/DecayFitNet.