File size: 1,471 Bytes
6a7b017
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
{
  "base_model_name": "Qwen/Qwen2.5-14B-Instruct",
  "router_embedding_model": "BAAI/bge-m3",
  "local_files_only": "False",
  "use_4bit": "True",
  "enable_real_training": "True",
  "enable_real_generation": "True",
  "enable_real_router_embedding": "True",
  "train_batch_size": "1",
  "gradient_accumulation_steps": "8",
  "learning_rate": "0.0002",
  "num_train_epochs": "1.0",
  "warmup_ratio": "0.03",
  "max_seq_length": "512",
  "max_new_tokens": "256",
  "train_max_examples": "1200",
  "lora_rank": "16",
  "lora_rank_fallback": "8",
  "lora_alpha": "32",
  "lora_dropout": "0.05",
  "router_confidence_threshold": "0.3",
  "router_hash_dim": "384",
  "min_samples_per_style": "100",
  "max_samples_per_style": "30000",
  "eval_sample_size": "10",
  "seed": "42",
  "data_dir": "/root/Multi-LoRA/data",
  "raw_data_dir": "/root/Multi-LoRA/data/raw",
  "processed_data_dir": "/root/Multi-LoRA/data/processed",
  "artifacts_dir": "/root/Multi-LoRA/artifacts",
  "adapters_dir": "/root/Multi-LoRA/artifacts/adapters",
  "persona_adapters_dir": "/root/Multi-LoRA/artifacts/adapters/persona",
  "router_dir": "/root/Multi-LoRA/artifacts/router",
  "logs_dir": "/root/Multi-LoRA/artifacts/logs",
  "reports_dir": "/root/Multi-LoRA/artifacts/reports",
  "processed_train_path": "/root/Multi-LoRA/data/processed/train.jsonl",
  "processed_eval_path": "/root/Multi-LoRA/data/processed/eval.jsonl",
  "processed_test_path": "/root/Multi-LoRA/data/processed/test.jsonl"
}