soumitsr commited on
Commit
d6a7b0e
·
verified ·
1 Parent(s): 4611c25

Model save

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -34,11 +34,11 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
- - train_batch_size: 8
38
- - eval_batch_size: 8
39
  - seed: 42
40
- - gradient_accumulation_steps: 8
41
- - total_train_batch_size: 64
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 1
@@ -51,6 +51,6 @@ The following hyperparameters were used during training:
51
  ### Framework versions
52
 
53
  - Transformers 4.51.3
54
- - Pytorch 2.6.0+cu124
55
  - Datasets 3.6.0
56
  - Tokenizers 0.21.1
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
+ - train_batch_size: 6
38
+ - eval_batch_size: 6
39
  - seed: 42
40
+ - gradient_accumulation_steps: 32
41
+ - total_train_batch_size: 192
42
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 1
 
51
  ### Framework versions
52
 
53
  - Transformers 4.51.3
54
+ - Pytorch 2.8.0.dev20250319+cu128
55
  - Datasets 3.6.0
56
  - Tokenizers 0.21.1