Qwen-Image-2512-Fun-Controlnet-Union

Github

Model Features

  • This ControlNet is added on 5 layer blocks. It supports multiple control conditionsโ€”including Canny, HED, Depth, Pose, MLSD and Scribble. It can be used like a standard ControlNet.
  • Inpainting mode is also supported.
  • When obtaining control images, acquiring them in a multi-resolution manner results in better generalization.
  • You can adjust control_context_scale for stronger control and better detail preservation. For better stability, we highly recommend using a detailed prompt. The optimal range for control_context_scale is from 0.70 to 0.95.

Results

Pose + Inpaint Output
Pose Output
Pose Output
Scribble Output
Canny Output
HED Output
Depth Output

Inference

Go to the VideoX-Fun repository for more details.

Please clone the VideoX-Fun repository and create the required directories:

# Clone the code
git clone https://github.com/aigc-apps/VideoX-Fun.git

# Enter VideoX-Fun's directory
cd VideoX-Fun

# Create model directories
mkdir -p models/Diffusion_Transformer
mkdir -p models/Personalized_Model

Then download the weights into models/Diffusion_Transformer and models/Personalized_Model.

๐Ÿ“ฆ models/
โ”œโ”€โ”€ ๐Ÿ“‚ Diffusion_Transformer/
โ”‚   โ””โ”€โ”€ ๐Ÿ“‚ Qwen-Image-2512/
โ”œโ”€โ”€ ๐Ÿ“‚ Personalized_Model/
โ”‚   โ””โ”€โ”€ ๐Ÿ“ฆ Qwen-Image-2512-Fun-Controlnet-Union.safetensors

Then run the file examples/qwenimage_fun/predict_t2i_control.py and examples/qwenimage_fun/predict_i2i_inpaint.py.

Downloads last month
1,226
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support