# modelblob.py from fastapi import APIRouter, Request from fastapi.responses import HTMLResponse, JSONResponse import json, os, pathlib router = APIRouter() LOCAL_BLOB_PATH = os.getenv("MODEL_BLOB_PATH", "/tmp/model_blob.json") # ---- EMBEDDED MODEL BLOB (exactly as provided) ---- EMBEDDED_MODEL_BLOB = { "name": "publishers/hf-misri/models/realismenginesdxl_v30vae", "versionId": "001", "openSourceCategory": "THIRD_PARTY_OWNED_OSS", "supportedActions": { "deploy": { "modelDisplayName": "misri/realismEngineSDXL_v30VAE", "containerSpec": { "imageUri": "us-docker.pkg.dev/deeplearning-platform-release/vertex-model-garden/hf-inference-toolkit.cu125.0-1.ubuntu2204.py311:model-garden.hf-inference-toolkit-0-1-release_20250927.00_p0", "env": [ {"name": "HF_TASK","value": "text-to-image"}, {"name": "MODEL_ID","value": "misri/realismEngineSDXL_v30VAE"}, {"name": "HF_MODEL_ID","value": "misri/realismEngineSDXL_v30VAE"}, {"name": "HF_REVISION","value": "7d2f2de544b4aa26148b3a16b3469ed6dbb38a5c"}, {"name": "DEPLOY_SOURCE","value": "UI_HF_VERIFIED_MODEL"} ], "ports": [{"containerPort": 8080}] }, "dedicatedResources": { "machineSpec": { "machineType": "a3-highgpu-1g", "acceleratorType": "NVIDIA_H100_80GB", "acceleratorCount": 1 }, "maxReplicaCount": 1 }, "deployTaskName": "1 NVIDIA_H100_80GB a3-highgpu-1g", "deployMetadata": {} }, "multiDeployVertex": { "multiDeployVertex": [ { "modelDisplayName": "misri/realismEngineSDXL_v30VAE", "containerSpec": { "imageUri": "us-docker.pkg.dev/deeplearning-platform-release/vertex-model-garden/hf-inference-toolkit.cu125.0-1.ubuntu2204.py311:model-garden.hf-inference-toolkit-0-1-release_20250927.00_p0", "env": [ {"name": "HF_TASK","value": "text-to-image"}, {"name": "MODEL_ID","value": "misri/realismEngineSDXL_v30VAE"}, {"name": "HF_MODEL_ID","value": "misri/realismEngineSDXL_v30VAE"}, {"name": "HF_REVISION","value": "7d2f2de544b4aa26148b3a16b3469ed6dbb38a5c"}, {"name": "DEPLOY_SOURCE","value": "UI_HF_VERIFIED_MODEL"} ], "ports": [{"containerPort": 8080}] }, "dedicatedResources": { "machineSpec": { "machineType": "a3-highgpu-1g", "acceleratorType": "NVIDIA_H100_80GB", "acceleratorCount": 1 }, "maxReplicaCount": 1 }, "deployTaskName": "1 NVIDIA_H100_80GB a3-highgpu-1g", "deployMetadata": {} }, { "modelDisplayName": "misri/realismEngineSDXL_v30VAE", "containerSpec": { "imageUri": "us-docker.pkg.dev/deeplearning-platform-release/vertex-model-garden/hf-inference-toolkit.cu125.0-1.ubuntu2204.py311:model-garden.hf-inference-toolkit-0-1-release_20250927.00_p0", "env": [ {"name": "HF_TASK","value": "text-to-image"}, {"name": "MODEL_ID","value": "misri/realismEngineSDXL_v30VAE"}, {"name": "HF_MODEL_ID","value": "misri/realismEngineSDXL_v30VAE"}, {"name": "HF_REVISION","value": "7d2f2de544b4aa26148b3a16b3469ed6dbb38a5c"}, {"name": "DEPLOY_SOURCE","value": "UI_HF_VERIFIED_MODEL"} ], "ports": [{"containerPort": 8080}] }, "dedicatedResources": { "machineSpec": { "machineType": "g2-standard-12", "acceleratorType": "NVIDIA_L4", "acceleratorCount": 1 }, "maxReplicaCount": 1 }, "deployTaskName": "1 NVIDIA_L4 g2-standard-12", "deployMetadata": {} } ] } } } # --------------------------------------------------- def _ensure_dir(p: str): pathlib.Path(p).parent.mkdir(parents=True, exist_ok=True) def _pretty(obj) -> str: try: return json.dumps(obj, indent=2) except Exception: return str(obj) @router.get("/modelblob", response_class=HTMLResponse) def view_model_blob(): """ Render: if /tmp/model_blob.json exists, show it; otherwise show the embedded blob. """ try: if os.path.exists(LOCAL_BLOB_PATH): raw = pathlib.Path(LOCAL_BLOB_PATH).read_text(encoding="utf-8") try: disp = json.dumps(json.loads(raw), indent=2) except json.JSONDecodeError: disp = raw source = f"File • {LOCAL_BLOB_PATH}" else: disp = _pretty(EMBEDDED_MODEL_BLOB) source = "Embedded (not yet written to file)" except Exception as e: disp = _pretty({"error": str(e)}) source = "Error" html = f"""