|
|
|
|
|
--- |
|
|
language: |
|
|
- ru |
|
|
- multilingual |
|
|
tags: |
|
|
- text-classification |
|
|
- topic-modeling |
|
|
- xlm-roberta |
|
|
--- |
|
|
|
|
|
# XLM-RoBERTa Base for Topic Classification |
|
|
|
|
|
Model fine-tuned for Russian social media topic classification. |
|
|
- Trained on: 512_posts_24_topics.csv |
|
|
- Epochs: 17 |
|
|
- Learning rate: 2e-05 |
|
|
|
|
|
| Epoch | Training Loss | Validation Loss | Accuracy | F1 Macro | F1 Weighted | |
|
|
|-------|---------------|-----------------|----------|----------|-------------| |
|
|
| 1 | No log | 3.026107 | 0.107527 | 0.018696 | 0.033548 | |
|
|
| 2 | No log | 2.870733 | 0.182796 | 0.052066 | 0.116982 | |
|
|
| 3 | No log | 2.696235 | 0.225806 | 0.062794 | 0.135901 | |
|
|
| 4 | No log | 2.468899 | 0.301075 | 0.117249 | 0.195119 | |
|
|
| 5 | No log | 2.275375 | 0.365591 | 0.164985 | 0.264671 | |
|
|
| 6 | No log | 2.103708 | 0.430108 | 0.246894 | 0.348294 | |
|
|
| 7 | No log | 1.953564 | 0.494624 | 0.297129 | 0.416058 | |
|
|
| 8 | No log | 1.905395 | 0.537634 | 0.356591 | 0.465515 | |
|
|
| 9 | No log | 1.783440 | 0.516129 | 0.349648 | 0.457348 | |
|
|
| 10 | No log | 1.733219 | 0.559140 | 0.385719 | 0.496481 | |
|
|
| 11 | No log | 1.668128 | 0.537634 | 0.361045 | 0.470044 | |
|
|
| 12 | No log | 1.633103 | 0.559140 | 0.387321 | 0.498477 | |
|
|
| 13 | No log | 1.588825 | 0.569892 | 0.392370 | 0.508810 | |
|
|
| 14 | No log | 1.581068 | 0.591398 | 0.410717 | 0.533581 | |
|
|
| 15 | No log | 1.560641 | 0.591398 | 0.413310 | 0.533612 | |
|
|
| 16 | No log | 1.545080 | 0.580645 | 0.403768 | 0.525466 | |
|
|
| 17 | No log | 1.550952 | 0.591398 | 0.410717 | 0.533581 | |