#!/usr/bin/env python3 from __future__ import annotations import json import os import subprocess import sys import time from http.server import BaseHTTPRequestHandler, HTTPServer from pathlib import Path from threading import Thread def _prepend_library_path(*paths: str) -> None: """Expose injected NVIDIA driver libraries before torch/triton imports.""" existing = [p for p in os.environ.get('LD_LIBRARY_PATH', '').split(':') if p] merged = [] for p in paths: if p and p not in merged: merged.append(p) for p in existing: if p not in merged: merged.append(p) os.environ['LD_LIBRARY_PATH'] = ':'.join(merged) _prepend_library_path( # HF Jobs injects the host driver under /usr/local/nvidia. Prefer that # over CUDA toolkit/compat libcuda stubs; using /usr/local/cuda/compat here # made A10G PyTorch report Error 803 despite nvidia-smi working. '/usr/local/nvidia/lib64', '/usr/local/nvidia/lib', '/usr/lib/x86_64-linux-gnu', ) # ============================================================================= # EARLY CUDA FABRIC MANAGER KICK (before ANY CUDA-touching imports) # ============================================================================= # On HF GPU hosts, cudaGetDeviceCount can transiently return not-ready errors # on first use. H200 fabric-manager is the worst case; A10G is usually ready # immediately, but the same early kick keeps the runtime deterministic. # synchronizes with the container's first driver call. Once any NVML/CUDA # call succeeds once (even just nvidia-smi), the fabric is up for the rest # of the container lifetime. # # Our previous approach (wait in a subprocess before training) didn't work # because the "initialization failed" state persisted across calls in the # same container. The real fix: kick the driver exactly once with # nvidia-smi, which is what successfully-working baseline containers do # implicitly via their first torch.cuda call. # # Must happen BEFORE `import torch` (because any import that eagerly calls # cudaGetDeviceCount will cache the Error 802 state). def _early_cuda_kick() -> None: deadline = time.time() + 120.0 attempt = 0 while time.time() < deadline: attempt += 1 r = subprocess.run(['nvidia-smi'], capture_output=True, text=True, timeout=30) if r.returncode == 0: gpu_line = next((ln.strip() for ln in (r.stdout or '').splitlines() if any(g in ln for g in ('A10', 'A100', 'H100', 'H200', 'RTX'))), 'gpu=unknown') print(f'[boot] nvidia-smi OK on attempt {attempt}: {gpu_line}', flush=True) break print(f'[boot] nvidia-smi attempt {attempt} rc={r.returncode} stderr={(r.stderr or "")[:120]}', flush=True) time.sleep(2) # After nvidia-smi, probe torch in a subprocess so any latent error state # doesn't leak into the main process's CUDA context. probe = 'import torch; import sys; sys.exit(0 if torch.cuda.is_available() else 1)' torch_deadline = time.time() + 120.0 t_attempt = 0 while time.time() < torch_deadline: t_attempt += 1 r = subprocess.run([sys.executable, '-c', probe], capture_output=True, text=True, timeout=60) if r.returncode == 0: print(f'[boot] torch.cuda.is_available() = True after {t_attempt} probe(s)', flush=True) return if t_attempt == 1: print(f'[boot] torch cuda probe {t_attempt}: {(r.stderr or "")[:200]}', flush=True) time.sleep(2) print('[boot] WARNING: torch.cuda never became ready — training will likely fail', flush=True) _early_cuda_kick() # Hydrate triton compilation cache from HF Hub before any triton/mamba_ssm import. # triton_cache_setup.py is copied next to this file by the job bash command. try: import triton_cache_setup as _tcs _tcs.setup() except ImportError: print('[boot] triton_cache_setup not found; skipping cache hydrate', flush=True) from huggingface_hub import HfApi # noqa: E402 (import after cuda kick) REPO_ROOT = Path('/workspace/feather') CACHE_ROOT = Path.home() / '.cache' / 'autoresearch' LOG_FILE = REPO_ROOT / 'run_domain_expanded.log' JOB_ID = os.environ.get('JOB_ID', 'local-job') OUTPUT_REPO = os.environ.get('HF_REPO_ID', 'icarus112/feather-pretrain-checkpoints') TOKEN = os.environ.get('HF_TOKEN') RUNTIME_MODE = os.environ.get('FEATHER_RUNTIME_MODE', 'space') APP_PORT = int(os.environ.get('PORT', '7860')) class _HealthHandler(BaseHTTPRequestHandler): def do_GET(self): if self.path in ('/', '/health', '/healthz', '/ready'): payload = { 'status': 'ok', 'mode': RUNTIME_MODE, 'job_id': JOB_ID, } body = json.dumps(payload).encode('utf-8') self.send_response(200) self.send_header('Content-Type', 'application/json') self.send_header('Content-Length', str(len(body))) self.end_headers() self.wfile.write(body) return self.send_response(404) self.end_headers() def log_message(self, format, *args): return def _start_health_server() -> HTTPServer: server = HTTPServer(('0.0.0.0', APP_PORT), _HealthHandler) thread = Thread(target=server.serve_forever, daemon=True) thread.start() print(f'[space] health server listening on 0.0.0.0:{APP_PORT}', flush=True) return server def upload_artifact(api: HfApi, path: Path, dest: str) -> None: if not path.exists(): print(f'[upload] skip missing {path}', flush=True) return api.upload_file( path_or_fileobj=str(path), path_in_repo=dest, repo_id=OUTPUT_REPO, repo_type='model', ) print(f'[upload] uploaded {path} -> {OUTPUT_REPO}/{dest}', flush=True) def _wait_for_cuda_ready(timeout_s: int = 120) -> None: """Block until CUDA is fully initialized or timeout. On H200 hosts with NVSwitch/fabric manager, nvidia driver setup can race with container start. cudaGetDeviceCount can return CUDA_ERROR_SYSTEM_NOT_READY (error 802) for the first few seconds, and any import that triggers @triton.autotune (e.g. mamba_ssm, torch amp utilities) blows up with "0 active drivers" if it happens during that window. We pre-init CUDA in a throwaway Python subprocess (so any error state does not leak into the main training process) and retry until torch.cuda reports ready. """ import time as _t probe = ( "import torch; " "import sys; " "avail = torch.cuda.is_available(); " "count = torch.cuda.device_count() if avail else 0; " "torch.empty(1, device='cuda') if (avail and count > 0) else None; " "from triton.runtime import driver; " "driver.active.get_current_device(); " "sys.exit(0 if (avail and count > 0) else 1)" ) deadline = _t.time() + timeout_s attempt = 0 while _t.time() < deadline: attempt += 1 r = subprocess.run(['python', '-c', probe], capture_output=True, text=True) if r.returncode == 0: print(f'[job] CUDA/Triton ready after {attempt} probe(s)', flush=True) return if attempt == 1: print(f'[job] CUDA not ready yet (will retry up to {timeout_s}s): {r.stderr.strip()[:200]}', flush=True) _t.sleep(2) print(f'[job] CUDA still not ready after {timeout_s}s — continuing anyway (training will likely fail)', flush=True) def run_job_mode() -> int: os.chdir(REPO_ROOT) os.environ.setdefault('HYDRA_TIME_BUDGET', '43200') os.environ.setdefault('HYDRA_TARGET_SHARDS', '2048') os.environ.setdefault('HYDRA_DOWNLOAD_WORKERS', '16') os.environ.setdefault('HYDRA_CKPT_INTERVAL', '1000') os.environ.setdefault('HYDRA_RESUME_CKPT', str(CACHE_ROOT / 'latest.pt')) os.environ.setdefault('FEATHER_GPU_PROFILE', 'a10g-large') os.environ.setdefault('HTM_CUDA_ARCH', 'sm_86') os.environ.setdefault('TORCH_CUDA_ARCH_LIST', '8.6') os.environ.setdefault('TRITON_CACHE_DIR', f"/workspace/triton_cache/{os.environ['FEATHER_GPU_PROFILE']}") os.environ.setdefault('TRITON_CACHE_REPO', f"icarus112/feather-triton-cache-{os.environ['FEATHER_GPU_PROFILE']}") print(f"[job] gpu_profile={os.environ['FEATHER_GPU_PROFILE']} htm_cuda_arch={os.environ['HTM_CUDA_ARCH']} torch_cuda_arch={os.environ['TORCH_CUDA_ARCH_LIST']}", flush=True) # CUDA readiness was kicked at module import via _early_cuda_kick. Keep # the wait as a second safety net — no-op if CUDA already ready. _wait_for_cuda_ready() cmd = [ 'bash', './scripts/run_domain_expanded_pretrain.sh', '--target-shards', os.environ['HYDRA_TARGET_SHARDS'], '--download-workers', os.environ['HYDRA_DOWNLOAD_WORKERS'], ] print('[job] ensuring retina.npz before training...', flush=True) try: sys.path.insert(0, str(REPO_ROOT)) from subsystems.sdr_retina import build_retina build_retina() except Exception as _retina_err: print(f'[job] retina bootstrap warning (train.py may still build it): {_retina_err}', flush=True) print('[job] starting Feather domain-expanded pretrain', flush=True) print(f'[job] command={cmd}', flush=True) proc = subprocess.run(cmd, check=False) # Push triton compilation cache back to HF Hub for next run. try: import triton_cache_setup as _tcs _tcs.teardown() except Exception as _tcs_err: print(f'[triton_cache] teardown error (non-fatal): {_tcs_err}', flush=True) if TOKEN: api = HfApi(token=TOKEN) try: api.create_repo(repo_id=OUTPUT_REPO, repo_type='model', private=True, exist_ok=True) except Exception as e: print(f'[upload] create_repo warning: {type(e).__name__}: {e}', flush=True) prefix = f'jobs/{JOB_ID}' try: upload_artifact(api, LOG_FILE, f'{prefix}/run_domain_expanded.log') upload_artifact(api, CACHE_ROOT / 'latest.pt', f'{prefix}/latest.pt') upload_artifact(api, CACHE_ROOT / 'pretrain_final.pt', f'{prefix}/pretrain_final.pt') except Exception as e: print(f'[upload] upload warning: {type(e).__name__}: {e}', flush=True) else: print('[upload] HF_TOKEN not set; skipping artifact upload', flush=True) return proc.returncode def run_space_mode() -> int: server = _start_health_server() print('[space] Feather runtime image ready', flush=True) try: while True: time.sleep(3600) finally: server.shutdown() server.server_close() def main() -> int: if RUNTIME_MODE == 'job': return run_job_mode() return run_space_mode() if __name__ == '__main__': raise SystemExit(main())