YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
Stable Diffusion v1.5 converted to LiteRT
This repository contains a LiteRT/TFLite export of the Hugging Face model stable-diffusion-v1-5/stable-diffusion-v1-5.
Base variants
fp32/: reference float export used byandroid-gpuandios-coremlint8/: mixed bundle with fp32 text encoder fallback, PT2E dynamic int8 UNet, and fp32 VAE fallback
Deployment profiles
android-qnn-npu: LiteRT Qualcomm AI Engine Direct (QNN) (android, preferred accelerator=NPU)android-gpu: LiteRT GPU delegate (android, preferred accelerator=GPU)android-cpu: LiteRT CPU/XNNPACK (android, preferred accelerator=CPU)ios-coreml: LiteRT Core ML delegate (ios, preferred accelerator=CORE_ML)
Profiles are emitted in conversion_manifest.json as manifest-level mappings onto the exported base variants. This avoids duplicating large model binaries while still letting each runtime pick backend-specific artifacts.
Files per exported base variant
text_encoder.tfliteunet.tflitevae_decoder.tflite
Shared assets
tokenizer/scheduler/configs/configs/text_encoder_runtime_config.jsonconversion_manifest.json
Notes
- Stable Diffusion v1.5 is a multi-stage pipeline, so this export is split into submodels.
- The notebook first tries to export the text encoder with INT32 token ids for better GPU/Core ML delegate compatibility and records the actual exported input dtype per variant and per deployment profile.
- The fp32 bundle is optional debug output; on CPU runtimes it is skipped by default to avoid kernel deaths during fp32 UNet conversion.
android-qnn-npuis a LiteRT/QNN-oriented deployment profile, not a Qualcomm AOT context binary.- Both exported base variants are smoke-tested by reloading the serialized LiteRT models and executing inference.
- The preview images in
preview/are decoder smoke tests, not final text-to-image samples.
- Downloads last month
- 209
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support