{ "cutoff_len": 512, "save_step": 1000, "train_lora_candidate_num": 2, "train_lora_simultaneously_num": 2, "train_strategy": "optim", "lora": [ { "name": "lora", "task_name": "", "optim": "adamw", "scheduler_type": "constant", "warmup_steps": 0, "lr": 2e-4, "batch_size": 16, "micro_batch_size": 8, "evaluate_batch_size": 16, "num_epochs": 2, "r": 80, "lora_alpha": 160, "lora_dropout": 0.05, "target_modules": { "q_proj": true, "k_proj": true, "v_proj": true, "o_proj": true, "gate_proj": true, "down_proj": true, "up_proj": true }, "group_by_length": false } ] }