Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: cc-by-sa-4.0
|
| 3 |
+
language:
|
| 4 |
+
- th
|
| 5 |
+
metrics:
|
| 6 |
+
- cer
|
| 7 |
+
- wer
|
| 8 |
+
library_name: espnet
|
| 9 |
+
pipeline_tag: automatic-speech-recognition
|
| 10 |
+
---
|
| 11 |
+
# Model Card for Model ID
|
| 12 |
+
|
| 13 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
+
|
| 15 |
+
This is the baseline model of Thai-central in [Thai-dialect corpus](https://github.com/SLSCU/thai-dialect-corpus).
|
| 16 |
+
|
| 17 |
+
The training recipe was based on wsj recipe in [espnet](https://github.com/espnet/espnet/).
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
### Model Description
|
| 22 |
+
|
| 23 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 24 |
+
|
| 25 |
+
This model is Hybrid CTC/Attention model with pre-trained HuBERT as the encoder.
|
| 26 |
+
|
| 27 |
+
This model trained on Thai-central for being the supervised pre-trained model in transfer-based curriculum learning experiment.
|
| 28 |
+
|
| 29 |
+
you can demo on colab with [this link](https://colab.research.google.com/drive/1stltGdpG9OV-sCl9QgkvEXZV7fGB2Ixe?usp=sharing). (Free google colab cannot inferences > 3 seconds of speech.)
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
## Evaluation
|
| 33 |
+
|
| 34 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 35 |
+
|
| 36 |
+
For evaluation, the metrics are CER and WER. before WER evaluation, transcriptions were re-tokenized using newmm tokenizer in [PyThaiNLP](https://github.com/PyThaiNLP/pythainlp)
|
| 37 |
+
|
| 38 |
+
In this reposirity, we also provide the vocabulary for building the newmm tokenizer using this script:
|
| 39 |
+
|
| 40 |
+
```python
|
| 41 |
+
from pythainlp import word_tokenize
|
| 42 |
+
|
| 43 |
+
tokenized_sentence_list = word_tokenize(<your_sentence>)
|
| 44 |
+
```
|
| 45 |
+
|
| 46 |
+
CER = 2.0
|
| 47 |
+
WER = 6.9
|
| 48 |
+
|
| 49 |
+
## Paper
|
| 50 |
+
|
| 51 |
+
[Thai Dialect Corpus and Transfer-based Curriculum Learning Investigation for Dialect Automatic Speech Recognition](https://www.isca-speech.org/archive/pdfs/interspeech_2023/suwanbandit23_interspeech.pdf)
|
| 52 |
+
```
|
| 53 |
+
@inproceedings{suwanbandit23_interspeech,
|
| 54 |
+
author={Artit Suwanbandit and Burin Naowarat and Orathai Sangpetch and Ekapol Chuangsuwanich},
|
| 55 |
+
title={{Thai Dialect Corpus and Transfer-based Curriculum Learning Investigation for Dialect Automatic Speech Recognition}},
|
| 56 |
+
year=2023,
|
| 57 |
+
booktitle={Proc. INTERSPEECH 2023},
|
| 58 |
+
pages={4069--4073},
|
| 59 |
+
doi={10.21437/Interspeech.2023-1828}
|
| 60 |
+
}
|
| 61 |
+
```
|