| lora_r: 32 | |
| lora_alpha: 16 | |
| lora_dropout: 0.05 | |
| lora_target_modules: | |
| lora_target_linear: true | |
| lora_fan_in_fan_out: | |
| gradient_accumulation_steps: 32 | |
| micro_batch_size: 2 | |
| num_epochs: 3 | |
| optimizer: paged_adamw_32bit | |
| lr_scheduler: cosine | |
| learning_rate: 0.0002 | |
| lora_r: 32 | |
| lora_alpha: 16 | |
| lora_dropout: 0.05 | |
| lora_target_modules: | |
| lora_target_linear: true | |
| lora_fan_in_fan_out: | |
| gradient_accumulation_steps: 32 | |
| micro_batch_size: 2 | |
| num_epochs: 3 | |
| optimizer: paged_adamw_32bit | |
| lr_scheduler: cosine | |
| learning_rate: 0.0002 | |