Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/srilaasya/asr_eval
Evaluating performance of a baseline Automatic Speech Recognition (ASR) model over 5 different speech/audio datasets.
https://github.com/srilaasya/asr_eval
Last synced: 2 days ago
JSON representation
Evaluating performance of a baseline Automatic Speech Recognition (ASR) model over 5 different speech/audio datasets.
- Host: GitHub
- URL: https://github.com/srilaasya/asr_eval
- Owner: srilaasya
- Created: 2023-04-27T15:15:37.000Z (almost 2 years ago)
- Default Branch: main
- Last Pushed: 2024-10-15T20:18:33.000Z (3 months ago)
- Last Synced: 2024-10-17T06:49:56.243Z (3 months ago)
- Language: Jupyter Notebook
- Size: 15.4 MB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# 🌱 ASR Wav2Vec2 model performance evaluation | [Access our final thesis here](https://drive.google.com/file/d/1UQ1UMkqdCXSJGFc7A-jQxpGEGhvopp3q/view?usp=sharing)
The aim of this study is to investigate the performance of the wav2vec model trained on five mutually exclusive subsets of the GigaSpeech dataset (a multi-domain English speech recognition corpus), each of which contains samples belonging to a specific category. Our goal is to train the model separately on each subset and evaluate its performance using a confusion matrix to derive conclusions about the generalization of the dataset to specific categories. The metric we plan to use is (w.e.r) word error rate. The wav2vec model has not previously been exposed to this dataset, and we seek to determine how it performs on these subsets. We describe the methodology used to create the subsets and preprocessed the data for training the model. We then evaluate the performance of the model on each subset separately and discuss the correlation between these categories. Our expectation is that the results will show that training the wav2vec model on these subsets separately can lead to improved performance and better generalization to the specific categories. Additionally, the confusion matrix analysis can provide insights into correlations between the different categories and if a combination of such data can improve the generalization of the model.
[Final working version colab notebook link](https://colab.research.google.com/drive/1hMwFGFsF5pJwqvCE7hyHA2kOw2T3pYgE?usp=sharing)
To run eval, after the training is done (after `trainer.train()` cell), move the `vocab.json` file into `/wav2vec2-base-speechcolab-demo/checkpoint-1000` folder. And then run eval.References:
[wav2vec2 documentation](https://huggingface.co/transformers/v4.8.2/model_doc/wav2vec2.html#transformers.Wav2Vec2Processor)
[wav2vec2 description HuggingFace](https://huggingface.co/facebook/wav2vec2-base)
[wav2vec2 fine-tuning example with timit](https://huggingface.co/blog/fine-tune-wav2vec2-english)
[wav2vec2 fine-tuning example with timit colab notebook](https://colab.research.google.com/drive/1FjTsqbYKphl9kL-eILgUc-bl4zVThL8F?usp=sharing#scrollTo=-np9xYK-wl8q)