Taja Kuzman
commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -216,7 +216,7 @@ and enriched with information which specific subtopics belong to the top-level t
|
|
| 216 |
## Training data
|
| 217 |
|
| 218 |
The model was fine-tuned on a training dataset consisting of 15,000 news in four languages (Croatian, Slovenian, Catalan and Greek).
|
| 219 |
-
The news texts were extracted from the [MaCoCu web corpora](
|
| 220 |
The training dataset was automatically annotated with the IPTC Media Topic labels by
|
| 221 |
the [GPT-4o](https://platform.openai.com/docs/models/gpt-4o) model (yielding 0.72 micro-F1 and 0.73 macro-F1 on the test dataset).
|
| 222 |
|
|
|
|
| 216 |
## Training data
|
| 217 |
|
| 218 |
The model was fine-tuned on a training dataset consisting of 15,000 news in four languages (Croatian, Slovenian, Catalan and Greek).
|
| 219 |
+
The news texts were extracted from the [MaCoCu-Genre web corpora](http://hdl.handle.net/11356/1969) based on the "News" genre label, predicted with the [X-GENRE classifier](https://huggingface.co/classla/xlm-roberta-base-multilingual-text-genre-classifier).
|
| 220 |
The training dataset was automatically annotated with the IPTC Media Topic labels by
|
| 221 |
the [GPT-4o](https://platform.openai.com/docs/models/gpt-4o) model (yielding 0.72 micro-F1 and 0.73 macro-F1 on the test dataset).
|
| 222 |
|