#!/usr/bin/env python3 from __future__ import annotations import json import os import subprocess import sys import time from http.server import BaseHTTPRequestHandler, HTTPServer from pathlib import Path from threading import Thread # ============================================================================= # EARLY CUDA FABRIC MANAGER KICK (before ANY CUDA-touching imports) # ============================================================================= # On H200 hosts, cudaGetDeviceCount can return Error 802 "system not yet # initialized" on first use, because nvidia-fabricmanager on the host # synchronizes with the container's first driver call. Once any NVML/CUDA # call succeeds once (even just nvidia-smi), the fabric is up for the rest # of the container lifetime. # # Our previous approach (wait in a subprocess before training) didn't work # because the "initialization failed" state persisted across calls in the # same container. The real fix: kick the driver exactly once with # nvidia-smi, which is what successfully-working baseline containers do # implicitly via their first torch.cuda call. # # Must happen BEFORE `import torch` (because any import that eagerly calls # cudaGetDeviceCount will cache the Error 802 state). def _early_cuda_kick() -> None: deadline = time.time() + 120.0 attempt = 0 while time.time() < deadline: attempt += 1 r = subprocess.run(['nvidia-smi'], capture_output=True, text=True, timeout=30) if r.returncode == 0 and 'H200' in (r.stdout or '') or 'H100' in (r.stdout or '') \ or 'A100' in (r.stdout or '') or r.returncode == 0: print(f'[boot] nvidia-smi OK on attempt {attempt}', flush=True) break print(f'[boot] nvidia-smi attempt {attempt} rc={r.returncode} stderr={(r.stderr or "")[:120]}', flush=True) time.sleep(2) # After nvidia-smi, probe torch in a subprocess so any latent error state # doesn't leak into the main process's CUDA context. probe = 'import torch; import sys; sys.exit(0 if torch.cuda.is_available() else 1)' torch_deadline = time.time() + 120.0 t_attempt = 0 while time.time() < torch_deadline: t_attempt += 1 r = subprocess.run([sys.executable, '-c', probe], capture_output=True, text=True, timeout=60) if r.returncode == 0: print(f'[boot] torch.cuda.is_available() = True after {t_attempt} probe(s)', flush=True) return if t_attempt == 1: print(f'[boot] torch cuda probe {t_attempt}: {(r.stderr or "")[:200]}', flush=True) time.sleep(2) print('[boot] WARNING: torch.cuda never became ready — training will likely fail', flush=True) _early_cuda_kick() # Hydrate triton compilation cache from HF Hub before any triton/mamba_ssm import. # triton_cache_setup.py is copied next to this file by the job bash command. try: import triton_cache_setup as _tcs _tcs.setup() except ImportError: print('[boot] triton_cache_setup not found; skipping cache hydrate', flush=True) from huggingface_hub import HfApi # noqa: E402 (import after cuda kick) REPO_ROOT = Path('/workspace/feather') CACHE_ROOT = Path.home() / '.cache' / 'autoresearch' LOG_FILE = REPO_ROOT / 'run_domain_expanded.log' JOB_ID = os.environ.get('JOB_ID', 'local-job') OUTPUT_REPO = os.environ.get('HF_REPO_ID', 'icarus112/feather-pretrain-checkpoints') TOKEN = os.environ.get('HF_TOKEN') RUNTIME_MODE = os.environ.get('FEATHER_RUNTIME_MODE', 'space') APP_PORT = int(os.environ.get('PORT', '7860')) class _HealthHandler(BaseHTTPRequestHandler): def do_GET(self): if self.path in ('/', '/health', '/healthz', '/ready'): payload = { 'status': 'ok', 'mode': RUNTIME_MODE, 'job_id': JOB_ID, } body = json.dumps(payload).encode('utf-8') self.send_response(200) self.send_header('Content-Type', 'application/json') self.send_header('Content-Length', str(len(body))) self.end_headers() self.wfile.write(body) return self.send_response(404) self.end_headers() def log_message(self, format, *args): return def _start_health_server() -> HTTPServer: server = HTTPServer(('0.0.0.0', APP_PORT), _HealthHandler) thread = Thread(target=server.serve_forever, daemon=True) thread.start() print(f'[space] health server listening on 0.0.0.0:{APP_PORT}', flush=True) return server def upload_artifact(api: HfApi, path: Path, dest: str) -> None: if not path.exists(): print(f'[upload] skip missing {path}', flush=True) return api.upload_file( path_or_fileobj=str(path), path_in_repo=dest, repo_id=OUTPUT_REPO, repo_type='model', ) print(f'[upload] uploaded {path} -> {OUTPUT_REPO}/{dest}', flush=True) def _wait_for_cuda_ready(timeout_s: int = 120) -> None: """Block until CUDA is fully initialized or timeout. On H200 hosts with NVSwitch/fabric manager, nvidia driver setup can race with container start. cudaGetDeviceCount can return CUDA_ERROR_SYSTEM_NOT_READY (error 802) for the first few seconds, and any import that triggers @triton.autotune (e.g. mamba_ssm, torch amp utilities) blows up with "0 active drivers" if it happens during that window. We pre-init CUDA in a throwaway Python subprocess (so any error state does not leak into the main training process) and retry until torch.cuda reports ready. """ import time as _t probe = ( "import torch; " "import sys; " "avail = torch.cuda.is_available(); " "count = torch.cuda.device_count() if avail else 0; " "sys.exit(0 if (avail and count > 0) else 1)" ) deadline = _t.time() + timeout_s attempt = 0 while _t.time() < deadline: attempt += 1 r = subprocess.run(['python', '-c', probe], capture_output=True, text=True) if r.returncode == 0: print(f'[job] CUDA ready after {attempt} probe(s)', flush=True) return if attempt == 1: print(f'[job] CUDA not ready yet (will retry up to {timeout_s}s): {r.stderr.strip()[:200]}', flush=True) _t.sleep(2) print(f'[job] CUDA still not ready after {timeout_s}s — continuing anyway (training will likely fail)', flush=True) def run_job_mode() -> int: os.chdir(REPO_ROOT) # Dynamic live patch from GitHub to bypass Space build errors GIT_REF = os.environ.get('FEATHER_GIT_REF') if GIT_REF and (REPO_ROOT / '.git').exists(): print(f'[bootstrap] dynamic sync to {GIT_REF}...', flush=True) subprocess.run(['git', 'fetch', 'origin'], cwd=REPO_ROOT, check=False) subprocess.run(['git', 'checkout', GIT_REF], cwd=REPO_ROOT, check=False) elif GIT_REF: print(f'[bootstrap] skipping dynamic sync (no .git in {REPO_ROOT})', flush=True) os.environ.setdefault('HYDRA_TIME_BUDGET', '43200') os.environ.setdefault('HYDRA_TARGET_SHARDS', '2048') os.environ.setdefault('HYDRA_DOWNLOAD_WORKERS', '16') os.environ.setdefault('HYDRA_CKPT_INTERVAL', '1000') os.environ.setdefault('HYDRA_RESUME_CKPT', str(CACHE_ROOT / 'latest.pt')) # CUDA readiness was kicked at module import via _early_cuda_kick. Keep # the wait as a second safety net — no-op if CUDA already ready. _wait_for_cuda_ready() cmd = [ 'bash', './scripts/run_domain_expanded_pretrain.sh', '--target-shards', os.environ['HYDRA_TARGET_SHARDS'], '--download-workers', os.environ['HYDRA_DOWNLOAD_WORKERS'], ] print('[job] starting Feather domain-expanded pretrain', flush=True) print(f'[job] command={cmd}', flush=True) proc = subprocess.run(cmd, check=False) # Push triton compilation cache back to HF Hub for next run. try: import triton_cache_setup as _tcs _tcs.teardown() except Exception as _tcs_err: print(f'[triton_cache] teardown error (non-fatal): {_tcs_err}', flush=True) if TOKEN: api = HfApi(token=TOKEN) try: api.create_repo(repo_id=OUTPUT_REPO, repo_type='model', private=True, exist_ok=True) except Exception as e: print(f'[upload] create_repo warning: {type(e).__name__}: {e}', flush=True) prefix = f'jobs/{JOB_ID}' try: upload_artifact(api, LOG_FILE, f'{prefix}/run_domain_expanded.log') upload_artifact(api, CACHE_ROOT / 'latest.pt', f'{prefix}/latest.pt') upload_artifact(api, CACHE_ROOT / 'pretrain_final.pt', f'{prefix}/pretrain_final.pt') except Exception as e: print(f'[upload] upload warning: {type(e).__name__}: {e}', flush=True) else: print('[upload] HF_TOKEN not set; skipping artifact upload', flush=True) return proc.returncode def run_space_mode() -> int: server = _start_health_server() print('[space] Feather runtime image ready', flush=True) try: while True: time.sleep(3600) finally: server.shutdown() server.server_close() def main() -> int: if RUNTIME_MODE == 'job': return run_job_mode() return run_space_mode() if __name__ == '__main__': raise SystemExit(main())