ltuzova commited on
Commit
7544b3e
·
1 Parent(s): 37f9d01

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -16
README.md CHANGED
@@ -16,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.8592
20
- - Accuracy: 0.7194
21
- - F1 Macro: 0.4819
22
 
23
  ## Model description
24
 
@@ -38,11 +38,11 @@ More information needed
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 2e-05
41
- - train_batch_size: 32
42
- - eval_batch_size: 64
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
- - total_train_batch_size: 64
46
  - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-06
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_ratio: 0.06
@@ -52,16 +52,14 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|
55
- | No log | 0.98 | 26 | 1.3259 | 0.5175 | 0.1137 |
56
- | No log | 2.0 | 53 | 1.1130 | 0.6228 | 0.2479 |
57
- | No log | 2.98 | 79 | 1.0243 | 0.6667 | 0.3126 |
58
- | 1.2194 | 4.0 | 106 | 0.9297 | 0.7018 | 0.3506 |
59
- | 1.2194 | 4.98 | 132 | 0.9334 | 0.7018 | 0.3593 |
60
- | 1.2194 | 6.0 | 159 | 0.8904 | 0.7368 | 0.5001 |
61
- | 1.2194 | 6.98 | 185 | 0.8714 | 0.7281 | 0.4661 |
62
- | 0.6526 | 8.0 | 212 | 0.8810 | 0.7368 | 0.4847 |
63
- | 0.6526 | 8.98 | 238 | 0.8807 | 0.7456 | 0.5552 |
64
- | 0.6526 | 9.81 | 260 | 0.8945 | 0.7193 | 0.5422 |
65
 
66
 
67
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.7271
20
+ - Accuracy: 0.7698
21
+ - F1 Macro: 0.6713
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 2e-05
41
+ - train_batch_size: 8
42
+ - eval_batch_size: 16
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
+ - total_train_batch_size: 16
46
  - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-06
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_ratio: 0.06
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro |
54
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|
55
+ | 1.419 | 1.0 | 105 | 1.0253 | 0.6842 | 0.3034 |
56
+ | 1.0312 | 2.0 | 211 | 0.9262 | 0.6842 | 0.3424 |
57
+ | 0.8072 | 3.0 | 316 | 0.8110 | 0.7018 | 0.3958 |
58
+ | 0.5688 | 4.0 | 422 | 0.7826 | 0.7632 | 0.6019 |
59
+ | 0.4064 | 5.0 | 527 | 0.7750 | 0.7719 | 0.6794 |
60
+ | 0.3165 | 6.0 | 633 | 0.8077 | 0.7544 | 0.6073 |
61
+ | 0.2172 | 7.0 | 738 | 0.9722 | 0.7544 | 0.6403 |
62
+ | 0.1455 | 8.0 | 844 | 0.9993 | 0.7719 | 0.6642 |
 
 
63
 
64
 
65
  ### Framework versions