Relacosm commited on
Commit
619e606
·
verified ·
1 Parent(s): dd550fa

Anti-overfitting 5-class sentiment model

Browse files
Files changed (1) hide show
  1. README.md +13 -12
README.md CHANGED
@@ -6,7 +6,6 @@ tags:
6
  - generated_from_trainer
7
  metrics:
8
  - accuracy
9
- - f1
10
  model-index:
11
  - name: mca-sentiment-analyzer-v2
12
  results: []
@@ -19,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-sentiment-latest](https://huggingface.co/cardiffnlp/twitter-roberta-base-sentiment-latest) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.9026
23
  - Accuracy: 1.0
24
- - F1: 1.0
25
  - F1 Macro: 1.0
 
26
 
27
  ## Model description
28
 
@@ -41,23 +40,25 @@ More information needed
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
- - learning_rate: 1e-05
45
- - train_batch_size: 8
46
- - eval_batch_size: 8
47
  - seed: 42
48
- - gradient_accumulation_steps: 2
49
  - total_train_batch_size: 16
50
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
  - lr_scheduler_type: linear
52
- - lr_scheduler_warmup_steps: 100
53
- - num_epochs: 5
54
  - mixed_precision_training: Native AMP
55
 
56
  ### Training results
57
 
58
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | F1 Macro |
59
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:---:|:--------:|
60
- | 1.284 | 3.88 | 50 | 0.9026 | 1.0 | 1.0 | 1.0 |
 
 
61
 
62
 
63
  ### Framework versions
 
6
  - generated_from_trainer
7
  metrics:
8
  - accuracy
 
9
  model-index:
10
  - name: mca-sentiment-analyzer-v2
11
  results: []
 
18
 
19
  This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-sentiment-latest](https://huggingface.co/cardiffnlp/twitter-roberta-base-sentiment-latest) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.0871
22
  - Accuracy: 1.0
 
23
  - F1 Macro: 1.0
24
+ - F1 Weighted: 1.0
25
 
26
  ## Model description
27
 
 
40
  ### Training hyperparameters
41
 
42
  The following hyperparameters were used during training:
43
+ - learning_rate: 5e-05
44
+ - train_batch_size: 4
45
+ - eval_batch_size: 4
46
  - seed: 42
47
+ - gradient_accumulation_steps: 4
48
  - total_train_batch_size: 16
49
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
50
  - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_ratio: 0.1
52
+ - num_epochs: 3
53
  - mixed_precision_training: Native AMP
54
 
55
  ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Macro | F1 Weighted |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:-----------:|
59
+ | 1.2197 | 0.8 | 20 | 0.9300 | 0.63 | 0.5436 | 0.5436 |
60
+ | 0.6318 | 1.6 | 40 | 0.3923 | 0.85 | 0.8255 | 0.8255 |
61
+ | 0.3146 | 2.4 | 60 | 0.0871 | 1.0 | 1.0 | 1.0 |
62
 
63
 
64
  ### Framework versions