Update README.md
Browse files
README.md
CHANGED
|
@@ -34,7 +34,10 @@ It is designed to support the development of open language models and for lingui
|
|
| 34 |
|
| 35 |
## Register labels and their abbreviations
|
| 36 |
|
| 37 |
-
Below is a list of the register labels predicted by the model. Note that some labels are hierarchical; when a sublabel is predicted, its parent label is also predicted.
|
|
|
|
|
|
|
|
|
|
| 38 |
|
| 39 |
- **MT:** Machine translated or generated
|
| 40 |
- **LY:** Lyrical
|
|
@@ -113,7 +116,7 @@ The model was trained using the Multilingual CORE Corpora, which will be publish
|
|
| 113 |
#### Training Hyperparameters
|
| 114 |
|
| 115 |
- **Batch size:** 8
|
| 116 |
-
- **Epochs:**
|
| 117 |
- **Learning rate:** 0.00005
|
| 118 |
- **Precision:** bfloat16 (non-mixed precision)
|
| 119 |
- **TF32:** Enabled
|
|
@@ -126,7 +129,19 @@ Average inference time (across 1000 iterations), using a single NVIDIA A100 GPU
|
|
| 126 |
|
| 127 |
## Evaluation
|
| 128 |
|
| 129 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 130 |
|
| 131 |
|
| 132 |
## Technical Specifications
|
|
|
|
| 34 |
|
| 35 |
## Register labels and their abbreviations
|
| 36 |
|
| 37 |
+
Below is a list of the register labels predicted by the model. Note that some labels are hierarchical; when a sublabel is predicted, its parent label is also predicted.
|
| 38 |
+
For a more detailed description, see [here](https://turkunlp.org/register-annotation-docs/).
|
| 39 |
+
|
| 40 |
+
The main labels are uppercase. To only include these main labels in the predictions, simply slice the model's output to keep only the uppercase labels.
|
| 41 |
|
| 42 |
- **MT:** Machine translated or generated
|
| 43 |
- **LY:** Lyrical
|
|
|
|
| 116 |
#### Training Hyperparameters
|
| 117 |
|
| 118 |
- **Batch size:** 8
|
| 119 |
+
- **Epochs:** 21
|
| 120 |
- **Learning rate:** 0.00005
|
| 121 |
- **Precision:** bfloat16 (non-mixed precision)
|
| 122 |
- **TF32:** Enabled
|
|
|
|
| 129 |
|
| 130 |
## Evaluation
|
| 131 |
|
| 132 |
+
**Evaluation results (micro-F1 for the languages the models was trained on):**
|
| 133 |
+
|
| 134 |
+
| Language | F1 (All labels) | F1 (Main labels) |
|
| 135 |
+
| -------- | --------------- | ---------------- |
|
| 136 |
+
| English | 0.72 |
|
| 137 |
+
| Finnish | 0.79 |
|
| 138 |
+
| French | 0.75 |
|
| 139 |
+
| Swedish | 0.81 |
|
| 140 |
+
| Turkish | 0.77 |
|
| 141 |
+
|
| 142 |
+
**Zero-shot evaluation results (micro-F1):**
|
| 143 |
+
|
| 144 |
+
|
| 145 |
|
| 146 |
|
| 147 |
## Technical Specifications
|