tarmomatic / README.md
jaahas's picture
Update README.md
0f11959 verified
---
tags:
- music
- video
---
All of the necessary models to run **[tarmomatic](https://codeberg.org/jaahas/tarmomatic)** local edition with ComfyUI.
You also need [ComfyUI-GGUF custom nodes](https://github.com/city96/ComfyUI-GGUF).
Place all of the models in their respective folders in the ComfyUI `models`-folder.
## Installation
1. Go to your ComfyUI directory
2. Download with HF CLI (or git):
```bash
curl -LsSf https://hf.co/cli/install.sh | bash
hf download jaahas/tarmomatic --local-dir models
```
## All Models List
| Model Filename | Required ComfyUI Folder | Used In |
|---|---|---|
| `flux1-schnell-fp8.safetensors` | `models/checkpoints` | Flux |
| `flux1-schnell-Q4_K_M.gguf` | `models/unet` | Flux |
| `qwen_2.5_vl_7b_fp8_scaled.safetensors` | `models/text_encoders` | Qwen Image Edit |
| `qwen_image_vae.safetensors` | `models/vae` | Qwen Image Edit |
| `Qwen-Image-Edit-2509-Lightning-4steps-V1.0-bf16.safetensors` | `models/loras` | Qwen Image Edit |
| `Qwen-Image-Edit-2509-Q4_K_M.gguf` | `models/unet` | Qwen Image Edit |
| `t5xxl_fp8_e4m3fn_scaled.safetensors` | `models/text_encoders` | LTX Video |
| `ltxv-2b-0.9.8-distilled-fp8.safetensors` | `models/checkpoints` | LTX Video |
| `umt5_xxl_fp8_e4m3fn_scaled.safetensors` | `models/text_encoders` | Wan Video |
| `wan2.2_vae.safetensors` | `models/vae` | Wan Video |
| `Wan2.2-TI2V-5B-Q5_K_M.gguf` | `models/unet` | Wan Video |
---
## Benchmarks (RTX 5090, cold start)
| Model | Speed |
|---|---|
| Flux Schnell Q4_K_M (1024x1024) | 28s |
| Qwen Image Edit 2509 Q4_K_M Lightning (1 image, 1024x1024) | 112s |
| Wan 2.2 TI2V 5B Q5_K_M (10s, 720p) | 460s |
| Wan 2.2 TI2V 5B Q5_K_M (10s, 720p, optimised) | 148s |
| LTXV 2b 0.9.8 distilled fp8 (10s, 512p) | 47s |
| TBA | --- |
| Wan 2.2 I2V A14B Q5_K_M Lightning (10s, 720p) | 1074s |
| Wan 2.2 I2V A14B Q5_K_M Lightning (10s, 480p) | 296s |
| Eigen Banana Qwen Image Edit 2509 Q4_K_M (1 image, 1024x1024) | 151s |
---
## Flux Models
Used for general image generation (Workflow: `flux_schnell-GGUF.json`).
- **`flux1-schnell-fp8.safetensors`**
- Folder: `models/checkpoints`
- **Note:** This provides the CLIP (Text Encoder) and VAE for the workflow.
- **`flux1-schnell-Q4_K_M.gguf`**
- Folder: `models/unet`
- **Note:** This provides the actual diffusion model (UNet) in a compressed (quantized) format for better performance.
## Qwen Image Models
Used for image editing and synthesis (Workflows: `image_qwen_image_edit_2509-GGUF-*.json`).
- **`qwen_2.5_vl_7b_fp8_scaled.safetensors`**
- Folder: `models/text_encoders`
- **`qwen_image_vae.safetensors`**
- Folder: `models/vae`
- **`Qwen-Image-Edit-2509-Lightning-4steps-V1.0-bf16.safetensors`**
- Folder: `models/loras`
- **`Qwen-Image-Edit-2509-Q4_K_M.gguf`**
- Folder: `models/unet`
## LTX Models
Used for image-to-video generation (Workflow: `ltxv_image_to_video.json`).
- **`t5xxl_fp8_e4m3fn_scaled.safetensors`**
- Folder: `models/text_encoders`
- **`ltxv-2b-0.9.8-distilled-fp8.safetensors`**
- Folder: `models/checkpoints`
## Wan Models
Used for video generation (Workflow: `video_wan2_2_5B_ti2v-GGUF.json`).
- **`umt5_xxl_fp8_e4m3fn_scaled.safetensors`**
- Folder: `models/text_encoders`
- **`wan2.2_vae.safetensors`**
- Folder: `models/vae`
- **`Wan2.2-TI2V-5B-Q5_K_M.gguf`**
- Folder: `models/unet`
---
## FAQ
### Why does Flux need both a GGUF and a Checkpoint?
The workflow uses a "hybrid" loading strategy:
1. **Checkpoint (`flux1-schnell-fp8.safetensors`):** Loads the **CLIP** (text understanding) and **VAE** (image decoding) components.
2. **GGUF (`flux1-schnell-Q4_K_M.gguf`):** Loads the **UNet** (image generation core).
This setup allows you to use a highly compressed, fast UNet (GGUF) while still getting the necessary support components from the standard checkpoint.