End of training
Browse files
README.md
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
base_model: xlm-roberta-large
|
| 4 |
+
tags:
|
| 5 |
+
- generated_from_trainer
|
| 6 |
+
metrics:
|
| 7 |
+
- f1
|
| 8 |
+
model-index:
|
| 9 |
+
- name: fine_tuned_xlm-roberta-large_2April
|
| 10 |
+
results: []
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 14 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 15 |
+
|
| 16 |
+
# fine_tuned_xlm-roberta-large_2April
|
| 17 |
+
|
| 18 |
+
This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the None dataset.
|
| 19 |
+
It achieves the following results on the evaluation set:
|
| 20 |
+
- Best F1: 75.9147
|
| 21 |
+
- Loss: 1.6634
|
| 22 |
+
- Exact: 39.3876
|
| 23 |
+
- F1: 56.7272
|
| 24 |
+
- Total: 3821
|
| 25 |
+
- Hasans Exact: 56.6152
|
| 26 |
+
- Hasans F1: 81.5886
|
| 27 |
+
- Hasans Total: 2653
|
| 28 |
+
- Noans Exact: 0.2568
|
| 29 |
+
- Noans F1: 0.2568
|
| 30 |
+
- Noans Total: 1168
|
| 31 |
+
- Best Exact: 60.4292
|
| 32 |
+
- Best Exact Thresh: 0.6508
|
| 33 |
+
- Best F1 Thresh: 0.9299
|
| 34 |
+
|
| 35 |
+
## Model description
|
| 36 |
+
|
| 37 |
+
More information needed
|
| 38 |
+
|
| 39 |
+
## Intended uses & limitations
|
| 40 |
+
|
| 41 |
+
More information needed
|
| 42 |
+
|
| 43 |
+
## Training and evaluation data
|
| 44 |
+
|
| 45 |
+
More information needed
|
| 46 |
+
|
| 47 |
+
## Training procedure
|
| 48 |
+
|
| 49 |
+
### Training hyperparameters
|
| 50 |
+
|
| 51 |
+
The following hyperparameters were used during training:
|
| 52 |
+
- learning_rate: 2e-05
|
| 53 |
+
- train_batch_size: 1
|
| 54 |
+
- eval_batch_size: 1
|
| 55 |
+
- seed: 42
|
| 56 |
+
- gradient_accumulation_steps: 16
|
| 57 |
+
- total_train_batch_size: 16
|
| 58 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 59 |
+
- lr_scheduler_type: linear
|
| 60 |
+
- num_epochs: 5
|
| 61 |
+
- mixed_precision_training: Native AMP
|
| 62 |
+
|
| 63 |
+
### Training results
|
| 64 |
+
|
| 65 |
+
| Training Loss | Epoch | Step | Best F1 | Validation Loss | Exact | F1 | Total | Hasans Exact | Hasans F1 | Hasans Total | Noans Exact | Noans F1 | Noans Total | Best Exact | Best Exact Thresh | Best F1 Thresh |
|
| 66 |
+
|:-------------:|:-----:|:----:|:-------:|:---------------:|:-------:|:-------:|:-----:|:------------:|:---------:|:------------:|:-----------:|:--------:|:-----------:|:----------:|:-----------------:|:--------------:|
|
| 67 |
+
| 2.0619 | 0.26 | 500 | 54.8058 | 1.4812 | 32.2952 | 51.4988 | 3821 | 46.5134 | 74.1714 | 2653 | 0.0 | 0.0 | 1168 | 41.4551 | 0.7820 | 0.8887 |
|
| 68 |
+
| 1.3598 | 0.53 | 1000 | 65.8410 | 1.2289 | 35.9068 | 54.1665 | 3821 | 51.7150 | 78.0136 | 2653 | 0.0 | 0.0 | 1168 | 51.7666 | 0.8149 | 0.8993 |
|
| 69 |
+
| 1.2347 | 0.79 | 1500 | 67.6083 | 1.1868 | 37.2416 | 55.0317 | 3821 | 53.6374 | 79.2597 | 2653 | 0.0 | 0.0 | 1168 | 53.7032 | 0.7981 | 0.8571 |
|
| 70 |
+
| 1.0961 | 1.05 | 2000 | 71.4132 | 1.1491 | 38.9165 | 55.8262 | 3821 | 56.0498 | 80.4041 | 2653 | 0.0 | 0.0 | 1168 | 57.5504 | 0.7733 | 0.8679 |
|
| 71 |
+
| 0.9003 | 1.32 | 2500 | 72.3291 | 1.2053 | 38.4193 | 56.0447 | 3821 | 55.3336 | 80.7188 | 2653 | 0.0 | 0.0 | 1168 | 57.5766 | 0.8557 | 0.9662 |
|
| 72 |
+
| 0.8705 | 1.58 | 3000 | 71.7222 | 1.1239 | 38.5239 | 56.1153 | 3821 | 55.4844 | 80.8204 | 2653 | 0.0 | 0.0 | 1168 | 56.6344 | 0.7408 | 0.8452 |
|
| 73 |
+
| 0.8655 | 1.84 | 3500 | 73.6273 | 1.0855 | 39.0212 | 56.4204 | 3821 | 56.2005 | 81.2599 | 2653 | 0.0 | 0.0 | 1168 | 58.7543 | 0.7574 | 0.8970 |
|
| 74 |
+
| 0.7431 | 2.11 | 4000 | 74.8323 | 1.1817 | 39.6231 | 56.5930 | 3821 | 56.9544 | 81.3954 | 2653 | 0.2568 | 0.2568 | 1168 | 59.7226 | 0.8032 | 0.9219 |
|
| 75 |
+
| 0.5738 | 2.37 | 4500 | 74.4675 | 1.2047 | 38.8642 | 56.6456 | 3821 | 55.8236 | 81.4334 | 2653 | 0.3425 | 0.3425 | 1168 | 58.8851 | 0.7003 | 0.8792 |
|
| 76 |
+
| 0.5904 | 2.63 | 5000 | 74.7345 | 1.1571 | 38.5763 | 56.4366 | 3821 | 55.5221 | 81.2455 | 2653 | 0.0856 | 0.0856 | 1168 | 59.1206 | 0.7922 | 0.8458 |
|
| 77 |
+
| 0.5831 | 2.89 | 5500 | 74.9378 | 1.1537 | 39.7278 | 56.6778 | 3821 | 57.2182 | 81.6306 | 2653 | 0.0 | 0.0 | 1168 | 59.9581 | 0.7947 | 0.8767 |
|
| 78 |
+
| 0.4785 | 3.16 | 6000 | 75.0234 | 1.3432 | 39.3353 | 56.8437 | 3821 | 56.1628 | 81.3795 | 2653 | 1.1130 | 1.1130 | 1168 | 59.2515 | 0.7999 | 0.8258 |
|
| 79 |
+
| 0.3774 | 3.42 | 6500 | 74.8641 | 1.4903 | 39.3876 | 56.6856 | 3821 | 56.2759 | 81.1894 | 2653 | 1.0274 | 1.0274 | 1168 | 59.6964 | 0.7004 | 0.9268 |
|
| 80 |
+
| 0.3882 | 3.68 | 7000 | 75.0504 | 1.3418 | 38.7857 | 56.4315 | 3821 | 55.8613 | 81.2759 | 2653 | 0.0 | 0.0 | 1168 | 59.4609 | 0.6782 | 0.9456 |
|
| 81 |
+
| 0.3747 | 3.95 | 7500 | 75.3181 | 1.3673 | 39.4399 | 56.6455 | 3821 | 56.8036 | 81.5840 | 2653 | 0.0 | 0.0 | 1168 | 59.6441 | 0.7541 | 0.9554 |
|
| 82 |
+
| 0.2748 | 4.21 | 8000 | 75.5835 | 1.5103 | 39.2829 | 56.6922 | 3821 | 56.5398 | 81.6136 | 2653 | 0.0856 | 0.0856 | 1168 | 60.1152 | 0.7237 | 0.9792 |
|
| 83 |
+
| 0.2346 | 4.47 | 8500 | 75.8283 | 1.6566 | 40.0157 | 57.2731 | 3821 | 57.0675 | 81.9225 | 2653 | 1.2842 | 1.2842 | 1168 | 60.5339 | 0.7154 | 0.9550 |
|
| 84 |
+
| 0.2339 | 4.74 | 9000 | 75.7051 | 1.6699 | 39.0474 | 56.5165 | 3821 | 56.0874 | 81.2475 | 2653 | 0.3425 | 0.3425 | 1168 | 60.3245 | 0.8887 | 0.9799 |
|
| 85 |
+
| 0.23 | 5.0 | 9500 | 75.9147 | 1.6634 | 39.3876 | 56.7272 | 3821 | 56.6152 | 81.5886 | 2653 | 0.2568 | 0.2568 | 1168 | 60.4292 | 0.6508 | 0.9299 |
|
| 86 |
+
|
| 87 |
+
|
| 88 |
+
### Framework versions
|
| 89 |
+
|
| 90 |
+
- Transformers 4.37.2
|
| 91 |
+
- Pytorch 1.13.1+cu117
|
| 92 |
+
- Datasets 2.16.1
|
| 93 |
+
- Tokenizers 0.15.2
|