Update README.md
Browse files
README.md
CHANGED
|
@@ -112,15 +112,15 @@ The model was fine-tuned using the Unsloth and LoRA.
|
|
| 112 |
|
| 113 |
#### Training Hyperparameters
|
| 114 |
|
| 115 |
-
|
| 116 |
-
|
| 117 |
-
|
| 118 |
-
|
| 119 |
-
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
|
| 125 |
#### Speeds, Sizes, Times [optional]
|
| 126 |
|
|
|
|
| 112 |
|
| 113 |
#### Training Hyperparameters
|
| 114 |
|
| 115 |
+
- batch size:8,
|
| 116 |
+
- gradient accumulation steps:4,
|
| 117 |
+
- optimizer:adamw_torch,
|
| 118 |
+
- learning rate:2e-5,
|
| 119 |
+
- warmup_steps:100,
|
| 120 |
+
- fp16:True,
|
| 121 |
+
- epochs:2,
|
| 122 |
+
- weight_decay:0.01,
|
| 123 |
+
- lr_scheduler_type:linear
|
| 124 |
|
| 125 |
#### Speeds, Sizes, Times [optional]
|
| 126 |
|