| model: | |
| name: meta-llama/Llama-3.1-8B-Instruct | |
| quantization: 4bit | |
| lora: | |
| rank: 16 | |
| alpha: 32 | |
| dropout: 0.05 | |
| target_modules: ["q_proj", "k_proj", "v_proj", "o_proj"] | |
| training: | |
| epochs: 3 | |
| batch_size: 2 | |
| gradient_accumulation_steps: 4 | |
| learning_rate: 2e-4 | |
| max_seq_length: 2048 | |
| warmup_ratio: 0.03 | |
| logging_steps: 10 | |
| save_steps: 100 | |
| output: | |
| base_dir: ./adapters | |