πŸ“¦ Model Artifacts

The quantized deployment artifacts for MedNTDs are publicly available.

We provide optimized formats for edge and cross-platform inference:

  • GGUF (llama.cpp compatible) – for high-performance CPU inference
  • TFLite (.task) – for mobile and embedded deployment

πŸ”— Hugging Face Repository

Model files are hosted on Hugging Face:
πŸ‘‰ https://huggingface.co/wlsgusjjn/MedNTDs

πŸ”— GitHub Repository

Full training pipeline, quantization scripts, and deployment code:
πŸ‘‰ https://github.com/wlsgusjjn/MedNTDs/

These artifacts include:

  • 4-bit quantized GGUF models for offline edge inference
  • LiteRT / TFLite task models for Flutter-based mobile integration
  • LoRA-adapted MedGemma checkpoints used in the 2-stage screening pipeline

All models are optimized for low-resource environments and designed for internet-independent deployment in rural clinical settings.

Downloads last month
8
GGUF
Model size
4B params
Architecture
gemma3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support