Update README.md
Browse files
README.md
CHANGED
|
@@ -9,4 +9,12 @@ license: apache-2.0
|
|
| 9 |
|
| 10 |
## Info
|
| 11 |
|
| 12 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
|
| 10 |
## Info
|
| 11 |
|
| 12 |
+
This Wav2Vec2 model was first pretrained on 500 hours Kalmyk TV recordings and 1000 hours Mongolian speech recognition dataset. After that, the model was finetuned on a 300 hours [Kalmyk synthetic STT dataset](https://github.com/tugstugi/mongolian-nlp#datasets) created by a voice conversion model.
|
| 13 |
+
* 50% WER on a private test set created from Kalmyk TV recordnings
|
| 14 |
+
* on clean voice recordings, the model should have much lower WER
|
| 15 |
+
* voice conversion info
|
| 16 |
+
* 300 hours [Kalmyk synthetic STT dataset](https://github.com/tugstugi/mongolian-nlp#datasets)
|
| 17 |
+
* The source voice is a Kalmyk female voice TTS
|
| 18 |
+
* Target voices are from the VCTK dataset
|
| 19 |
+
* example data: https://twitter.com/tugstugi/status/1409111296897912835
|
| 20 |
+
* each WAV has a different text created from Kalmyk books
|