https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500
https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500
Last synced: about 2 months ago
JSON representation
- Host: GitHub
- URL: https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500
- Owner: csukuangfj
- Created: 2021-10-14T12:58:53.000Z (over 3 years ago)
- Default Branch: master
- Last Pushed: 2023-11-09T10:59:23.000Z (over 1 year ago)
- Last Synced: 2025-02-08T08:12:55.464Z (4 months ago)
- Size: 686 KB
- Stars: 2
- Watchers: 2
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
## How to clone this repo
**Caution**: `git clone https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500.git`
won't work. If you use it, you will encounter the following error in [icefall][icefall]:```
_pickle.UnpicklingError: invalid load key, 'v'
```Please use the following commands to clone this repo:
```
sudo apt-get install git-lfs
git clone https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500.git
cd icefall-asr-conformer-ctc-bpe-500
git lfs pull
```## Description
This repo contains a pre-trained model using [icefall][icefall]
with the following command (it uses 7 GPUs):```
cd egs/librispeech/ASR
./conformer_ctc/train.py \
--world-size 7 \
--max-duration 150 \
--bucketing-sampler 1 \
--full-libri 1 \
--start-epoch 0 \
--num-epochs 50 \
--exp-dir conformer_ctc/exp \
--lang-dir data/lang_bpe_500
```The model `exp/pretrained.pt` is obtained by the following command:
```
cd egs/librispeech/ASR
./conformer_ctc/export.py \
--epoch 49 \
--avg 15 \
--exp-dir conformer_ctc/exp \
--lang-dir data/lang_bpe_500 \
--jit 0
```You can use `exp/pretrained.pt` to compute the WER for `test-clean` and `test-other`:
```
cd egs/librispeech/ASR
cd conformer_ctc/exp
ln -s pretrained.pt epoch-99.pt
cd ../..
./conformer_ctc/decode.py \
--exp-dir ./conformer_ctc/exp \
--lang-dir ./data/lang_bpe_500 \
--epoch 99 \
--avg 1 \
--max-duration 30 \
--nbest-scale 0.5
```You will get the following log:
```
2021-10-14 20:01:39,246 INFO [decode.py:474] batch 0/804, cuts processed until now is 6
2021-10-14 20:03:19,789 INFO [decode.py:474] batch 100/804, cuts processed until now is 398
2021-10-14 20:04:53,371 INFO [decode.py:474] batch 200/804, cuts processed until now is 792
2021-10-14 20:06:22,611 INFO [decode.py:474] batch 300/804, cuts processed until now is 1160
2021-10-14 20:07:50,482 INFO [decode.py:474] batch 400/804, cuts processed until now is 1574
2021-10-14 20:09:20,129 INFO [decode.py:474] batch 500/804, cuts processed until now is 1965
2021-10-14 20:10:53,932 INFO [decode.py:474] batch 600/804, cuts processed until now is 2308
2021-10-14 20:12:36,852 INFO [decode.py:474] batch 700/804, cuts processed until now is 2498
2021-10-14 20:14:26,477 INFO [decode.py:474] batch 800/804, cuts processed until now is 2614
2021-10-14 20:16:16,519 INFO [decode.py:523]
For test-clean, WER of different settings are:
ngram_lm_scale_1.1_attention_scale_0.9 2.56 best for test-clean
ngram_lm_scale_1.1_attention_scale_1.0 2.56
ngram_lm_scale_1.2_attention_scale_1.0 2.56
ngram_lm_scale_0.9_attention_scale_0.6 2.57
ngram_lm_scale_0.9_attention_scale_0.7 2.57
ngram_lm_scale_1.0_attention_scale_0.6 2.57
ngram_lm_scale_1.0_attention_scale_0.9 2.57
ngram_lm_scale_1.1_attention_scale_0.7 2.57
ngram_lm_scale_1.1_attention_scale_1.1 2.57
ngram_lm_scale_1.2_attention_scale_0.7 2.57
ngram_lm_scale_1.2_attention_scale_0.9 2.57
ngram_lm_scale_1.2_attention_scale_1.1 2.57
ngram_lm_scale_1.2_attention_scale_1.2 2.57
ngram_lm_scale_1.3_attention_scale_0.9 2.57
ngram_lm_scale_1.3_attention_scale_1.1 2.57
ngram_lm_scale_1.5_attention_scale_1.0 2.57
ngram_lm_scale_1.5_attention_scale_1.1 2.57
ngram_lm_scale_1.5_attention_scale_1.2 2.57
ngram_lm_scale_1.5_attention_scale_1.3 2.57
ngram_lm_scale_2.0_attention_scale_1.7 2.572021-10-14 20:16:17,979 INFO [decode.py:474] batch 0/782, cuts processed until now is 6
2021-10-14 20:17:54,037 INFO [decode.py:474] batch 100/782, cuts processed until now is 434
2021-10-14 20:19:26,335 INFO [decode.py:474] batch 200/782, cuts processed until now is 885
2021-10-14 20:20:52,910 INFO [decode.py:474] batch 300/782, cuts processed until now is 1327
2021-10-14 20:22:15,968 INFO [decode.py:474] batch 400/782, cuts processed until now is 1807
2021-10-14 20:23:42,595 INFO [decode.py:474] batch 500/782, cuts processed until now is 2238
2021-10-14 20:25:06,502 INFO [decode.py:474] batch 600/782, cuts processed until now is 2584
2021-10-14 20:26:46,868 INFO [decode.py:474] batch 700/782, cuts processed until now is 2785
2021-10-14 20:30:11,556 INFO [decode.py:523]
For test-other, WER of different settings are:
ngram_lm_scale_1.5_attention_scale_1.9 5.8 best for test-other
ngram_lm_scale_1.7_attention_scale_1.9 5.8
ngram_lm_scale_1.7_attention_scale_2.0 5.8
ngram_lm_scale_1.9_attention_scale_2.0 5.8
ngram_lm_scale_1.3_attention_scale_1.5 5.81
ngram_lm_scale_1.5_attention_scale_1.5 5.81
ngram_lm_scale_1.5_attention_scale_1.7 5.81
ngram_lm_scale_1.7_attention_scale_1.7 5.81
ngram_lm_scale_1.5_attention_scale_1.3 5.82
ngram_lm_scale_1.9_attention_scale_1.9 5.82
ngram_lm_scale_1.2_attention_scale_1.2 5.83
ngram_lm_scale_1.3_attention_scale_1.3 5.83
ngram_lm_scale_1.3_attention_scale_1.7 5.83
ngram_lm_scale_1.5_attention_scale_1.2 5.83
ngram_lm_scale_1.5_attention_scale_2.0 5.83
ngram_lm_scale_2.0_attention_scale_2.0 5.83
```[icefall]: https://github.com/k2-fsa/icefall
### Note
This repo uses `git lfs`. See