| { |
| "base_model_name": "Qwen/Qwen2.5-14B-Instruct", |
| "router_embedding_model": "BAAI/bge-m3", |
| "local_files_only": "False", |
| "use_4bit": "True", |
| "enable_real_training": "True", |
| "enable_real_generation": "True", |
| "enable_real_router_embedding": "True", |
| "train_batch_size": "1", |
| "gradient_accumulation_steps": "8", |
| "learning_rate": "0.0002", |
| "num_train_epochs": "1.0", |
| "warmup_ratio": "0.03", |
| "max_seq_length": "512", |
| "max_new_tokens": "256", |
| "train_max_examples": "1200", |
| "lora_rank": "16", |
| "lora_rank_fallback": "8", |
| "lora_alpha": "32", |
| "lora_dropout": "0.05", |
| "router_confidence_threshold": "0.3", |
| "router_hash_dim": "384", |
| "min_samples_per_style": "100", |
| "max_samples_per_style": "30000", |
| "eval_sample_size": "10", |
| "seed": "42", |
| "data_dir": "/root/Multi-LoRA/data", |
| "raw_data_dir": "/root/Multi-LoRA/data/raw", |
| "processed_data_dir": "/root/Multi-LoRA/data/processed", |
| "artifacts_dir": "/root/Multi-LoRA/artifacts", |
| "adapters_dir": "/root/Multi-LoRA/artifacts/adapters", |
| "persona_adapters_dir": "/root/Multi-LoRA/artifacts/adapters/persona", |
| "router_dir": "/root/Multi-LoRA/artifacts/router", |
| "logs_dir": "/root/Multi-LoRA/artifacts/logs", |
| "reports_dir": "/root/Multi-LoRA/artifacts/reports", |
| "processed_train_path": "/root/Multi-LoRA/data/processed/train.jsonl", |
| "processed_eval_path": "/root/Multi-LoRA/data/processed/eval.jsonl", |
| "processed_test_path": "/root/Multi-LoRA/data/processed/test.jsonl" |
| } |