https://github.com/fightnyy/distillbart_eck
Repository for distillation of English, Chinese, Korean Multilingual BART
https://github.com/fightnyy/distillbart_eck
chinese distillation english korean mbart multilingual
Last synced: 6 months ago
JSON representation
Repository for distillation of English, Chinese, Korean Multilingual BART
- Host: GitHub
- URL: https://github.com/fightnyy/distillbart_eck
- Owner: fightnyy
- Created: 2021-04-05T12:37:08.000Z (over 4 years ago)
- Default Branch: master
- Last Pushed: 2021-07-31T05:48:52.000Z (about 4 years ago)
- Last Synced: 2025-03-25T17:49:23.735Z (7 months ago)
- Topics: chinese, distillation, english, korean, mbart, multilingual
- Language: Python
- Homepage:
- Size: 45.9 KB
- Stars: 6
- Watchers: 1
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
## Distill mbart For ECK
* Among mbart, I fine-tuned the models for English, Chinese, and Korean.
* Each model has a different number of layers. (For example, if model denotes 12-3, it represent this model is composed of __12 Encoder__ and __3 Decoder__
* We plan to develop OpenSource called OpenSFT soon. Please look forward to it!| __BLEU__ \ Model | 12-3 English | 12 -3 Korean | 12 - 3 Chinese |
| ---------------- | ------------ | ------------ | -------------- |
| __1st epoch__ | __53__ | __35__ | __27__ |
| 2nd epoch | 52 | 35 | 25 |
| 3rd epoch | 51 | 33 | 23 |__Inference Time :__ 0.5S
__Parameter Size__ : 262M(3.15G)
| __BLEU__ \ Model | 9-3 English | 9 -3 Korean | 9 - 3 Chinese |
| ---------------- | ----------- | ----------- | ------------- |
| 1st epoch | 54 | __36__ | 24 |
| 2nd epoch | __55__ | 35 | __25__ |
| 3rd epoch | 54 | 35 | 23 |__Inference Time :__ 0.2S
__Parameter Size__ : 224M(2.7G)