David Cole
commited on
Upload training_config.yml with huggingface_hub
Browse files- training_config.yml +3 -3
training_config.yml
CHANGED
|
@@ -6,7 +6,7 @@ model:
|
|
| 6 |
apply_lora_to_mlp: false
|
| 7 |
apply_lora_to_output: false
|
| 8 |
lora_rank: 32
|
| 9 |
-
lora_alpha:
|
| 10 |
perception_tokens: 2
|
| 11 |
use_clip: false
|
| 12 |
tokenizer:
|
|
@@ -46,7 +46,7 @@ batch_size: 4
|
|
| 46 |
optimizer:
|
| 47 |
_component_: torch.optim.AdamW
|
| 48 |
weight_decay: 0.001
|
| 49 |
-
lr: 1e-
|
| 50 |
betas:
|
| 51 |
- 0.9
|
| 52 |
- 0.998
|
|
@@ -59,7 +59,7 @@ grad_clip:
|
|
| 59 |
_component_: torch.nn.utils.clip_grad_norm
|
| 60 |
max_norm: 2.0
|
| 61 |
norm_type: 2
|
| 62 |
-
epochs:
|
| 63 |
max_steps_per_epoch: null
|
| 64 |
gradient_accumulation_steps: 32
|
| 65 |
compile: true
|
|
|
|
| 6 |
apply_lora_to_mlp: false
|
| 7 |
apply_lora_to_output: false
|
| 8 |
lora_rank: 32
|
| 9 |
+
lora_alpha: 32
|
| 10 |
perception_tokens: 2
|
| 11 |
use_clip: false
|
| 12 |
tokenizer:
|
|
|
|
| 46 |
optimizer:
|
| 47 |
_component_: torch.optim.AdamW
|
| 48 |
weight_decay: 0.001
|
| 49 |
+
lr: 1e-7
|
| 50 |
betas:
|
| 51 |
- 0.9
|
| 52 |
- 0.998
|
|
|
|
| 59 |
_component_: torch.nn.utils.clip_grad_norm
|
| 60 |
max_norm: 2.0
|
| 61 |
norm_type: 2
|
| 62 |
+
epochs: 8
|
| 63 |
max_steps_per_epoch: null
|
| 64 |
gradient_accumulation_steps: 32
|
| 65 |
compile: true
|