MedGemma_StructCore / README.md
DocUA's picture
Fix HF Space metadata colorTo
04dfd71

A newer version of the Gradio SDK is available: 6.8.0

Upgrade
metadata
title: MedGemma StructCore Demo
emoji: 🩺
colorFrom: blue
colorTo: indigo
sdk: gradio
python_version: '3.10'
app_file: app.py
pinned: false

MedGemma StructCore Demo (HF Spaces Zero)

This directory contains deployment assets for Hugging Face Spaces Zero.

What is included

  • app.py: Space entrypoint for the StructCore demo UI.
  • requirements.txt: minimal dependencies for this demo.

Recommended deployment flow

Use the packaging script from the repository root:

bash scripts/prepare_hf_zero_challenge_space.sh

It creates a ready-to-push bundle in:

.dist/hf_zero_challenge_demo_space/

Then push that bundle to your HF Space repository.

Model repository (two-stage)

Target model repo:

  • https://huggingface.co/DocUA/medgemma-1.5-4b-it-gguf-q5-k-m-two-stage

Upload/update Stage1 and Stage2 artifacts from this project repo:

python3 scripts/hf_upload_two_stage_models.py \
  --repo-id DocUA/medgemma-1.5-4b-it-gguf-q5-k-m-two-stage \
  --stage1-file /absolute/path/to/stage1.gguf \
  --stage2-file /absolute/path/to/stage2.gguf \
  --stage1-path-in-repo stage1/medgemma-stage1-q5_k_m.gguf \
  --stage2-path-in-repo stage2/medgemma-stage2-q5_k_m.gguf

Requires HF_TOKEN with write access to the model repo.

Space runtime configuration

Set these variables/secrets in the HF Space settings:

  • STRUCTCORE_BACKEND_MODE=pipeline (or mock as safe default)
  • STRUCTCORE_STAGE1_URL=<your_openai_compat_stage1_url>
  • STRUCTCORE_STAGE1_MODEL=<model_alias_from_stage1_/v1/models>
  • STRUCTCORE_STAGE2_URL=<your_openai_compat_stage2_url>
  • STRUCTCORE_STAGE2_MODEL=<model_alias_from_stage2_/v1/models>

Important:

  • Space itself does not serve GGUF automatically from the model repo.
  • GGUF files in HF model repo are the source-of-truth artifacts.
  • Actual inference in pipeline mode requires reachable OpenAI-compatible endpoints running those artifacts.