# Fast fine-tune for Zenith - World's First Autonomous AI Development Partner param( [string]$PythonExe = "python" ) Write-Host "🚀 Setting up ZENITH fine-tune for Aspetos (<2h training)..." Write-Host "Building the most advanced AI development partner in existence!" # Env vars for stable training (adjust if needed) $Env:BASE_MODEL = "D:\aspetos\DeepSeek-Coder-V2-Lite-Instruct" $Env:OUTPUT_DIR = "outputs\zenith-lora" $Env:DATA_PATH = "data\zenith_combined.jsonl" $Env:EPOCHS = "1" $Env:BATCH = "4" # Balanced for A100 $Env:GRAD_ACC = "4" # Effective batch size = 16 $Env:LR = "1e-4" # Stable learning rate for proper convergence $Env:STEPS = "180" # Sufficient steps for learning $Env:MAX_SEQ_LEN = "2048" $Env:USE_4BIT = "1" $Env:SEED = "42" $Env:MAX_GRAD_NORM = "1.0" # Gradient clipping $Env:WEIGHT_DECAY = "0.01" $Env:WARMUP_RATIO = "0.05" $Env:EARLY_STOP_PATIENCE = "5" # Allow more patience for learning $Env:EVAL_STEPS = "40" # Balanced evaluation frequency $Env:SAVE_STEPS = "40" # Balanced save frequency Write-Host "Installing dependencies..." & $PythonExe -m pip install -r requirements.txt if ($LASTEXITCODE -ne 0) { throw "pip install failed" } Write-Host "🎯 Starting ZENITH fine-tune (target: 200 steps, ~1-2 hours on GPU)..." Write-Host "Training the world's most advanced autonomous development partner!" & $PythonExe train.py if ($LASTEXITCODE -ne 0) { throw "Training failed" } Write-Host "✅ ZENITH training complete - Ready for deployment!"