--- pipeline_tag: any-to-any library_name: transformers tags: - text-to-image - image-editing - image-understanding - vision-language - multimodal - unified-model license: mit --- ## 🌌 Unipic3-DMD-Model(Distribution Matching Distillation)
Skywork Logo

GitHub Repo GitHub Stars GitHub Forks

## 📖 Introduction
Model Teaser
**UniPic3-DMD-Model** is a few-step image editing and multi-image composition model trained using **Distribution Matching Distillation (DMD)**. The model directly matches the **output distribution of a high-quality teacher model**, enabling sharp, visually detailed generations in very few inference steps. It is designed to maximize **perceptual quality and realism**, closely imitating strong proprietary or large teacher models. This model is initialized from a consistency-trained checkpoint and further refined via distribution-level distillation. ## 📊 Benchmarks
Model Teaser
## 🧠 Usage ### 1. Clone the Repository ```bash git clone https://github.com/SkyworkAI/UniPic cd UniPic-3 ``` ### 2. Set Up the Environment ```bash conda create -n unipic python=3.10 conda activate unipic3 pip install -r requirements.txt ``` ### 3.Batch Inference ```bash transformer_path = "Skywork/Unipic3-DMD/ema_transformer" python -m torch.distributed.launch --nproc_per_node=1 --master_port 29501 --use_env \ qwen_image_edit_fast/batch_inference.py \ --jsonl_path data/val.jsonl \ --output_dir work_dirs/output \ --distributed \ --num_inference_steps 8 \ --true_cfg_scale 4.0 \ --transformer transformer_path \ --skip_existing ``` ## 📄 License This model is released under the MIT License. ## Citation If you use Skywork UniPic 3.0 in your research, please cite: ``` @article{wei2026skywork, title={Skywork UniPic 3.0: Unified Multi-Image Composition via Sequence Modeling}, author={Wei, Hongyang and Liu, Hongbo and Wang, Zidong and Peng, Yi and Xu, Baixin and Wu, Size and Zhang, Xuying and He, Xianglong and Liu, Zexiang and Wang, Peiyu and others}, journal={arXiv preprint arXiv:2601.15664}, year={2026} } ```