Update README.md
Browse files
README.md
CHANGED
|
@@ -10,7 +10,7 @@ tags:
|
|
| 10 |
- pictograms
|
| 11 |
- translation
|
| 12 |
metrics:
|
| 13 |
-
-
|
| 14 |
inference: false
|
| 15 |
---
|
| 16 |
|
|
@@ -60,7 +60,7 @@ fairseq-train \
|
|
| 60 |
|
| 61 |
### Evaluation
|
| 62 |
|
| 63 |
-
The model was evaluated with
|
| 64 |
|
| 65 |
```bash
|
| 66 |
fairseq-generate exp_commonvoice/data-bin/commonvoice.tokenized.fr-frp \
|
|
@@ -82,8 +82,8 @@ Generate test with beam=5: BLEU4 = 82.60, 92.5/85.5/79.5/74.1 (BP=1.000, ratio=1
|
|
| 82 |
Comparison to other translation models :
|
| 83 |
| **Model** | **validation** | **test** |
|
| 84 |
|:-----------:|:-----------------------:|:-----------------------:|
|
| 85 |
-
|
|
| 86 |
-
| t2p-nmt-commonvoice | 86.0 | 82.6 |
|
| 87 |
| t2p-mbart-large-cc25-commonvoice | 72.3 | 72.3 |
|
| 88 |
| t2p-nllb-200-distilled-600M-commonvoice | **87.4** | **87.6** |
|
| 89 |
|
|
|
|
| 10 |
- pictograms
|
| 11 |
- translation
|
| 12 |
metrics:
|
| 13 |
+
- sacrebleu
|
| 14 |
inference: false
|
| 15 |
---
|
| 16 |
|
|
|
|
| 60 |
|
| 61 |
### Evaluation
|
| 62 |
|
| 63 |
+
The model was evaluated with sacreBLEU, where we compared the reference pictogram translation with the model hypothesis.
|
| 64 |
|
| 65 |
```bash
|
| 66 |
fairseq-generate exp_commonvoice/data-bin/commonvoice.tokenized.fr-frp \
|
|
|
|
| 82 |
Comparison to other translation models :
|
| 83 |
| **Model** | **validation** | **test** |
|
| 84 |
|:-----------:|:-----------------------:|:-----------------------:|
|
| 85 |
+
| t2p-t5-large-commonvoice | 86.3 | 86.5 |
|
| 86 |
+
| **t2p-nmt-commonvoice** | 86.0 | 82.6 |
|
| 87 |
| t2p-mbart-large-cc25-commonvoice | 72.3 | 72.3 |
|
| 88 |
| t2p-nllb-200-distilled-600M-commonvoice | **87.4** | **87.6** |
|
| 89 |
|