c-ho commited on
Commit
3d6f0b3
·
verified ·
1 Parent(s): b07cda8

c-ho/academic_main_text_classifier_de

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -21,11 +21,11 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.2835
25
- - Accuracy: 0.9319
26
- - Precision: 0.9319
27
- - Recall: 0.9319
28
- - F1: 0.9319
29
 
30
  ## Model description
31
 
@@ -44,8 +44,8 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 5e-05
48
- - train_batch_size: 32
49
  - eval_batch_size: 64
50
  - seed: 42
51
  - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
@@ -57,11 +57,11 @@ The following hyperparameters were used during training:
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
59
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
60
- | 0.8491 | 1.0 | 268 | 0.3697 | 0.8858 | 0.8858 | 0.8858 | 0.8858 |
61
- | 0.4033 | 2.0 | 536 | 0.3948 | 0.8946 | 0.8946 | 0.8946 | 0.8946 |
62
- | 0.2694 | 3.0 | 804 | 0.3032 | 0.9193 | 0.9193 | 0.9193 | 0.9193 |
63
- | 0.1998 | 4.0 | 1072 | 0.2782 | 0.9315 | 0.9315 | 0.9315 | 0.9315 |
64
- | 0.1549 | 5.0 | 1340 | 0.2835 | 0.9319 | 0.9319 | 0.9319 | 0.9319 |
65
 
66
 
67
  ### Framework versions
 
21
 
22
  This model is a fine-tuned version of [FacebookAI/xlm-roberta-base](https://huggingface.co/FacebookAI/xlm-roberta-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.2560
25
+ - Accuracy: 0.9245
26
+ - Precision: 0.9245
27
+ - Recall: 0.9245
28
+ - F1: 0.9245
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - learning_rate: 2e-05
48
+ - train_batch_size: 64
49
  - eval_batch_size: 64
50
  - seed: 42
51
  - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
 
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
59
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
60
+ | No log | 1.0 | 134 | 0.6173 | 0.7627 | 0.7627 | 0.7627 | 0.7627 |
61
+ | 0.9527 | 2.0 | 268 | 0.4141 | 0.8811 | 0.8811 | 0.8811 | 0.8811 |
62
+ | 0.4191 | 3.0 | 402 | 0.3220 | 0.8965 | 0.8965 | 0.8965 | 0.8965 |
63
+ | 0.4191 | 4.0 | 536 | 0.2663 | 0.9175 | 0.9175 | 0.9175 | 0.9175 |
64
+ | 0.2567 | 5.0 | 670 | 0.2560 | 0.9245 | 0.9245 | 0.9245 | 0.9245 |
65
 
66
 
67
  ### Framework versions