dtorber commited on
Commit
85f1df9
·
verified ·
1 Parent(s): 56360dc

Model save

Browse files
Files changed (1) hide show
  1. README.md +18 -22
README.md CHANGED
@@ -18,15 +18,14 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [google-bert/bert-base-cased](https://huggingface.co/google-bert/bert-base-cased) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.2458
22
- - F1 Macro: 0.8817
23
- - F1: 0.9130
24
- - F1 Neg: 0.8503
25
- - Acc: 0.89
26
- - Prec: 0.9130
27
- - Recall: 0.9130
28
- - Mcc: 0.7634
29
- - Millor Epoca: 6
30
 
31
  ## Model description
32
 
@@ -49,28 +48,25 @@ The following hyperparameters were used during training:
49
  - train_batch_size: 8
50
  - eval_batch_size: 8
51
  - seed: 42
52
- - distributed_type: multi-GPU
53
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
  - lr_scheduler_type: linear
55
- - num_epochs: 7
56
  - mixed_precision_training: Native AMP
57
 
58
  ### Training results
59
 
60
- | Training Loss | Epoch | Step | Validation Loss | F1 Macro | F1 | F1 Neg | Acc | Prec | Recall | Mcc | Millor Epoca |
61
- |:-------------:|:-----:|:-----:|:---------------:|:--------:|:------:|:------:|:------:|:------:|:------:|:------:|:------------:|
62
- | 0.2675 | 1.0 | 3801 | 0.5655 | 0.8736 | 0.9115 | 0.8357 | 0.885 | 0.8876 | 0.9368 | 0.7498 | 1 |
63
- | 0.1016 | 2.0 | 7602 | 0.7133 | 0.8640 | 0.8926 | 0.8354 | 0.87 | 0.9351 | 0.8538 | 0.7337 | 1 |
64
- | 0.041 | 3.0 | 11403 | 1.0065 | 0.8702 | 0.9101 | 0.8303 | 0.8825 | 0.8815 | 0.9407 | 0.7442 | 1 |
65
- | 0.0211 | 4.0 | 15204 | 1.2972 | 0.8736 | 0.9115 | 0.8357 | 0.885 | 0.8876 | 0.9368 | 0.7498 | 1 |
66
- | 0.0108 | 5.0 | 19005 | 1.3592 | 0.8677 | 0.9080 | 0.8273 | 0.88 | 0.8810 | 0.9368 | 0.7387 | 1 |
67
- | 0.002 | 6.0 | 22806 | 1.2458 | 0.8817 | 0.9130 | 0.8503 | 0.89 | 0.9130 | 0.9130 | 0.7634 | 6 |
68
- | 0.0 | 7.0 | 26607 | 1.3670 | 0.8766 | 0.9133 | 0.8399 | 0.8875 | 0.8910 | 0.9368 | 0.7554 | 6 |
69
 
70
 
71
  ### Framework versions
72
 
73
- - Transformers 4.39.3
74
  - Pytorch 2.3.0+cu121
75
  - Datasets 2.18.0
76
- - Tokenizers 0.15.2
 
18
 
19
  This model is a fine-tuned version of [google-bert/bert-base-cased](https://huggingface.co/google-bert/bert-base-cased) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.4624
22
+ - F1 Macro: 0.9016
23
+ - F1: 0.9343
24
+ - F1 Neg: 0.8689
25
+ - Acc: 0.9125
26
+ - Prec: 0.9432
27
+ - Recall: 0.9257
28
+ - Mcc: 0.8036
 
29
 
30
  ## Model description
31
 
 
48
  - train_batch_size: 8
49
  - eval_batch_size: 8
50
  - seed: 42
 
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 5
54
  - mixed_precision_training: Native AMP
55
 
56
  ### Training results
57
 
58
+ | Training Loss | Epoch | Step | Validation Loss | F1 Macro | F1 | F1 Neg | Acc | Prec | Recall | Mcc |
59
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:------:|:------:|:------:|:------:|:------:|
60
+ | No log | 1.0 | 474 | 0.3773 | 0.8286 | 0.8676 | 0.7896 | 0.8375 | 0.9064 | 0.8320 | 0.6623 |
61
+ | 0.4662 | 2.0 | 948 | 0.4197 | 0.8638 | 0.9070 | 0.8205 | 0.8775 | 0.8819 | 0.9336 | 0.7305 |
62
+ | 0.3045 | 3.0 | 1422 | 0.4950 | 0.8808 | 0.9219 | 0.8397 | 0.895 | 0.8794 | 0.9688 | 0.7711 |
63
+ | 0.2105 | 4.0 | 1896 | 0.4999 | 0.8817 | 0.9175 | 0.8459 | 0.8925 | 0.9019 | 0.9336 | 0.7644 |
64
+ | 0.1246 | 5.0 | 2370 | 0.5317 | 0.8889 | 0.9198 | 0.8581 | 0.8975 | 0.9216 | 0.9180 | 0.7779 |
 
 
65
 
66
 
67
  ### Framework versions
68
 
69
+ - Transformers 4.40.1
70
  - Pytorch 2.3.0+cu121
71
  - Datasets 2.18.0
72
+ - Tokenizers 0.19.1