Spaces:
Sleeping
Sleeping
A newer version of the Gradio SDK is available:
6.8.0
metadata
title: MedGemma StructCore Demo
emoji: 🩺
colorFrom: blue
colorTo: indigo
sdk: gradio
python_version: '3.10'
app_file: app.py
pinned: false
MedGemma StructCore Demo (HF Spaces Zero)
This directory contains deployment assets for Hugging Face Spaces Zero.
What is included
app.py: Space entrypoint for the StructCore demo UI.requirements.txt: minimal dependencies for this demo.
Recommended deployment flow
Use the packaging script from the repository root:
bash scripts/prepare_hf_zero_challenge_space.sh
It creates a ready-to-push bundle in:
.dist/hf_zero_challenge_demo_space/
Then push that bundle to your HF Space repository.
Model repository (two-stage)
Target model repo:
https://huggingface.co/DocUA/medgemma-1.5-4b-it-gguf-q5-k-m-two-stage
Upload/update Stage1 and Stage2 artifacts from this project repo:
python3 scripts/hf_upload_two_stage_models.py \
--repo-id DocUA/medgemma-1.5-4b-it-gguf-q5-k-m-two-stage \
--stage1-file /absolute/path/to/stage1.gguf \
--stage2-file /absolute/path/to/stage2.gguf \
--stage1-path-in-repo stage1/medgemma-stage1-q5_k_m.gguf \
--stage2-path-in-repo stage2/medgemma-stage2-q5_k_m.gguf
Requires HF_TOKEN with write access to the model repo.
Space runtime configuration
Set these variables/secrets in the HF Space settings:
STRUCTCORE_BACKEND_MODE=pipeline(ormockas safe default)STRUCTCORE_STAGE1_URL=<your_openai_compat_stage1_url>STRUCTCORE_STAGE1_MODEL=<model_alias_from_stage1_/v1/models>STRUCTCORE_STAGE2_URL=<your_openai_compat_stage2_url>STRUCTCORE_STAGE2_MODEL=<model_alias_from_stage2_/v1/models>
Important:
- Space itself does not serve GGUF automatically from the model repo.
- GGUF files in HF model repo are the source-of-truth artifacts.
- Actual inference in
pipelinemode requires reachable OpenAI-compatible endpoints running those artifacts.