FiveC commited on
Commit
d253dbf
·
verified ·
1 Parent(s): 9628985

End of training

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [IAmSkyDra/BARTBana_v5](https://huggingface.co/IAmSkyDra/BARTBana_v5) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.8556
22
- - Sacrebleu: 1.5240
23
 
24
  ## Model description
25
 
@@ -39,10 +39,10 @@ More information needed
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 2e-05
42
- - train_batch_size: 64
43
- - eval_batch_size: 64
44
  - seed: 42
45
- - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
  - num_epochs: 3
48
  - mixed_precision_training: Native AMP
@@ -51,14 +51,14 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Sacrebleu |
53
  |:-------------:|:-----:|:-----:|:---------------:|:---------:|
54
- | 0.3694 | 1.0 | 15848 | 2.5203 | 1.0776 |
55
- | 0.2075 | 2.0 | 31696 | 2.7832 | 1.3857 |
56
- | 0.1651 | 3.0 | 47544 | 2.8556 | 1.5240 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - Transformers 4.57.3
62
- - Pytorch 2.9.1+cu128
63
- - Datasets 4.4.2
64
  - Tokenizers 0.22.1
 
18
 
19
  This model is a fine-tuned version of [IAmSkyDra/BARTBana_v5](https://huggingface.co/IAmSkyDra/BARTBana_v5) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 2.7448
22
+ - Sacrebleu: 1.3613
23
 
24
  ## Model description
25
 
 
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 2e-05
42
+ - train_batch_size: 128
43
+ - eval_batch_size: 128
44
  - seed: 42
45
+ - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
  - num_epochs: 3
48
  - mixed_precision_training: Native AMP
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Sacrebleu |
53
  |:-------------:|:-----:|:-----:|:---------------:|:---------:|
54
+ | 0.4874 | 1.0 | 8322 | 2.4353 | 0.9260 |
55
+ | 0.2892 | 2.0 | 16644 | 2.6754 | 1.2124 |
56
+ | 0.2431 | 3.0 | 24966 | 2.7448 | 1.3613 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - Transformers 4.57.3
62
+ - Pytorch 2.9.0+cu126
63
+ - Datasets 4.0.0
64
  - Tokenizers 0.22.1