--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - kontext - web-design - ui-ux - design-transformation - instruction-following base_model: black-forest-labs/FLUX.1-dev pipeline_tag: image-to-image --- # InstructDesign-Kontext Fine-tuned FLUX.1 Kontext [dev] model for transforming web interfaces with natural language instructions. ## Model Description InstructDesign-Kontext is a LoRA fine-tuned version of FLUX.1 Kontext [dev] specialized in: - Web interface transformations - UI/UX redesigns - Style transfers (dark mode, glassmorphism, etc.) - Device mockup generation - Design system applications ### Training Details - **Base Model**: FLUX.1 Kontext [dev] - **Training Method**: LoRA fine-tuning - **Dataset**: 937 web interface transformation pairs - **Training Steps**: 10,000 - **Training Time**: ~2 weeks (iterative) ## File Structure ``` ├── unet/ │ └── flux1-kontext-dev.safetensors (23GB) ├── loras/ │ └── flux_kontext_lora_v4_consolidated_000010000.safetensors (2.6GB) ├── clip/ │ ├── clip_l.safetensors (235MB) │ └── t5xxl_fp8_e4m3fn_scaled.safetensors (4.9GB) ├── vae/ │ └── ae.safetensors (320MB) └── workflow/ └── workflow_api.json ``` ## Usage with ComfyUI ### 1. Download Models Download all model files to your ComfyUI models directory: ```bash # Main model ComfyUI/models/diffusion_models/flux1-kontext-dev.safetensors # LoRA ComfyUI/models/loras/flux_kontext_lora_v4_consolidated_000010000.safetensors # CLIP encoders ComfyUI/models/clip/clip_l.safetensors ComfyUI/models/clip/t5xxl_fp8_e4m3fn_scaled.safetensors # VAE ComfyUI/models/vae/ae.safetensors ``` ### 2. Load Workflow Import the `workflow/workflow_api.json` into ComfyUI. ### 3. Configure Parameters Default optimal settings: - **Steps**: 20 - **CFG**: 1.0 - **Guidance**: 5.0 - **Sampler**: er_sde - **Scheduler**: kl_optimal - **LoRA Strength**: 1.0 ## API Usage See the [GitHub repository](https://github.com/tercumantanumut/instructdesign-kontext) for full API implementation. ```python # Example API request import requests response = requests.post("http://localhost:8000/api/generate", json={ "positive_prompt": "Transform this interface to dark mode with purple accents", "input_image": "website.png", "seed": 42, "steps": 20, "cfg": 1.0, "guidance": 5.0 }) ``` ## Transformation Capabilities ### Style Transformations - Dark/light mode conversion - Glassmorphism effects - Neumorphic design - Material Design application - iOS/Android style transfer ### UI/UX Improvements - Layout modernization - Accessibility enhancements - Color scheme updates - Typography improvements - Interactive element additions ### Mockup Generation - Device frame placement - Environment contextualization - Billboard/poster mockups - Multi-device presentations ## Example Prompts ``` "Transform this website to dark mode with purple gradient accents" "Apply glassmorphism effect with blurred backgrounds" "Convert to material design 3 with rounded corners" "Place this interface on an iPhone 15 Pro in a coffee shop" "Modernize this dated interface with current design trends" ``` ## Performance - **Inference Time**: ~45-70 seconds (NVIDIA GPU 16GB+) - **VRAM Requirements**: 16GB minimum, 24GB recommended - **Quality**: 85%+ instruction adherence rate ## Docker Deployment Complete Docker setup available in the GitHub repository with: - ComfyUI container - FastAPI wrapper - Queue management - Auto-scaling workers ## Limitations - Requires significant VRAM (16GB+) - Best results with clear, specific instructions - Limited to single image transformations (no batch in single inference) ## Citation If you use this model, please cite: ```bibtex @misc{instructdesign-kontext, title={InstructDesign-Kontext: Web Interface Transformation with FLUX}, author={tercumantanumut}, year={2024}, publisher={Hugging Face} } ``` ## License This model is subject to the FLUX.1 [dev] Non-Commercial License. ## Acknowledgments - FLUX.1 team for the base model - ComfyUI for the inference framework - 10 weeks of iterative development