Model-52 / qlora.yml
FlippedAIChat's picture
Upload folder using huggingface_hub
c415577 verified
raw
history blame contribute delete
250 Bytes
lora_r: 32
lora_alpha: 16
lora_dropout: 0.05
lora_target_modules:
lora_target_linear: true
lora_fan_in_fan_out:
gradient_accumulation_steps: 32
micro_batch_size: 2
num_epochs: 3
optimizer: paged_adamw_32bit
lr_scheduler: cosine
learning_rate: 0.0002