SanDiegoDude commited on
Commit
f5e38fa
·
verified ·
1 Parent(s): 491e92d

Update README with inference tool instructions

Browse files
Files changed (1) hide show
  1. README.md +31 -5
README.md CHANGED
@@ -30,13 +30,39 @@ The full bf16 transformer is 32.5 GB — too large for consumer GPUs like the RT
30
 
31
  RTX 4090 (Ada Lovelace, SM89) has native FP8 hardware support.
32
 
33
- ## Usage
34
 
35
- This is a drop-in replacement for the original safetensors checkpoint. The model code automatically detects FP8 tensors and upcasts them to bf16 at load time.
36
 
37
- 1. Download this repo into `ckpts_infer/transformer/`
38
- 2. Ensure you also have the VAE and text encoder from the [original repo](https://huggingface.co/jdopensource/JoyAI-Image-Edit) or [safetensors conversion](https://huggingface.co/SanDiegoDude/JoyAI-Image-Edit-Safetensors)
39
- 3. Run inference normally — the loading code handles the FP8→bf16 upcast transparently
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  ## Conversion
42
 
 
30
 
31
  RTX 4090 (Ada Lovelace, SM89) has native FP8 hardware support.
32
 
33
+ ## Inference Tool
34
 
35
+ A Gradio UI, CLI, and REST API for running inference are available at **[SanDiegoDude/JoyAI-Image](https://github.com/SanDiegoDude/JoyAI-Image)** a handy way to run the model until proper ComfyUI integration lands. Features include auto-download from HuggingFace, multiple memory modes, and bitsandbytes quantization for the text encoder.
36
 
37
+ ### Quick start
38
+
39
+ ```bash
40
+ git clone https://github.com/SanDiegoDude/JoyAI-Image.git
41
+ cd JoyAI-Image
42
+ python -m venv .venv && source .venv/bin/activate
43
+ pip install -e .
44
+
45
+ # Models auto-download on first run (FP8 transformer is the default)
46
+ # Default: FP8 transformer + 8-bit text encoder, offload mode
47
+ python app.py
48
+
49
+ # Minimum VRAM (~13 GB active, fits RTX 4090)
50
+ python app.py --nf4-dit --4bit-vlm
51
+
52
+ # CLI inference
53
+ python inference.py \
54
+ --prompt "Turn the plate blue" \
55
+ --image test_images/test_1.jpg \
56
+ --output result.png \
57
+ --steps 18 --guidance-scale 4.0 --seed 42
58
+
59
+ # Headless REST API (with optional ComfyUI connector node)
60
+ python app.py --headless-api 7500
61
+ ```
62
+
63
+ ## Also available
64
+
65
+ - **[Full bf16 safetensors](https://huggingface.co/SanDiegoDude/JoyAI-Image-Edit-Safetensors)** — original precision, also the source weights for runtime NF4 quantization (~8 GB)
66
 
67
  ## Conversion
68