File size: 404 Bytes
ed1b365 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 | model:
name: meta-llama/Llama-3.1-8B-Instruct
quantization: 4bit
lora:
rank: 16
alpha: 32
dropout: 0.05
target_modules: ["q_proj", "k_proj", "v_proj", "o_proj"]
training:
epochs: 3
batch_size: 2
gradient_accumulation_steps: 4
learning_rate: 2e-4
max_seq_length: 2048
warmup_ratio: 0.03
logging_steps: 10
save_steps: 100
output:
base_dir: ./adapters
|