{ "cutoff_len": 512, "save_step": 1000, "train_lora_candidate_num": 2, "train_lora_simultaneously_num": 2, "train_strategy": "optim", "lora": [ { "name": "loramoe", "task_name": "", "optim": "adamw", "scheduler_type": "constant", "warmup_steps": 0, "lr": 2e-4, "batch_size": 16, "micro_batch_size": 8, "evaluate_batch_size": 16, "num_epochs": 2, "r": 22, "lora_alpha": 44, "lora_dropout": 0.05, "target_modules": { "q_proj": false, "k_proj": false, "v_proj": false, "dense": false, "fc1": true, "fc2": true }, "routing_strategy": "loramoe", "num_experts": 6, "group_by_length": false } ] }