s8n29 commited on
Commit
e199f8a
·
verified ·
1 Parent(s): caeb059

End of training

Browse files
Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -16,9 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [unsloth/gemma-2-9b-it-bnb-4bit](https://huggingface.co/unsloth/gemma-2-9b-it-bnb-4bit) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 2.7063
20
- - Acc: 0.45
21
- - Log Loss: 2.7062
22
 
23
  ## Model description
24
 
@@ -46,14 +46,18 @@ The following hyperparameters were used during training:
46
  - optimizer: Use adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_steps: 20
49
- - num_epochs: 1
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Acc | Log Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|:----:|:--------:|
56
- | 3.5673 | 1.0 | 10 | 2.7063 | 0.45 | 2.7062 |
 
 
 
 
57
 
58
 
59
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [unsloth/gemma-2-9b-it-bnb-4bit](https://huggingface.co/unsloth/gemma-2-9b-it-bnb-4bit) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 1.3774
20
+ - Acc: 0.35
21
+ - Log Loss: 1.3774
22
 
23
  ## Model description
24
 
 
46
  - optimizer: Use adamw_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_steps: 20
49
+ - num_epochs: 5
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Acc | Log Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|:----:|:--------:|
56
+ | 3.0058 | 1.0 | 10 | 1.8528 | 0.4 | 1.8527 |
57
+ | 2.9092 | 2.0 | 20 | 1.7207 | 0.45 | 1.7206 |
58
+ | 2.6161 | 3.0 | 30 | 1.5368 | 0.4 | 1.5367 |
59
+ | 2.3656 | 4.0 | 40 | 1.4192 | 0.35 | 1.4191 |
60
+ | 2.1831 | 5.0 | 50 | 1.3774 | 0.35 | 1.3774 |
61
 
62
 
63
  ### Framework versions