Improve model card: Add metadata, links, and usage example
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,5 +1,74 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
pipeline_tag: image-to-image
|
| 4 |
+
library_name: transformers
|
| 5 |
---
|
| 6 |
|
| 7 |
+
# UniREdit-Bagel
|
| 8 |
+
|
| 9 |
+
This repository contains **UniREdit-Bagel**, a model developed for **UniREditBench: A Unified Reasoning-based Image Editing Benchmark**. The paper introduces UniREditBench as a unified benchmark for reasoning-based image editing evaluation and UniREdit-Bagel as a model fine-tuned on the UniREdit-Data-100K dataset, demonstrating substantial improvements in both in-domain and out-of-distribution settings.
|
| 10 |
+
|
| 11 |
+
* [**Paper on arXiv**](https://arxiv.org/abs/2511.01295)
|
| 12 |
+
* [**Project Page**](https://maplebb.github.io/UniREditBench/)
|
| 13 |
+
* [**GitHub Repository**](https://github.com/Maplebb/UniREditBench)
|
| 14 |
+
|
| 15 |
+
## Introduction
|
| 16 |
+
|
| 17 |
+
We propose **UniREditBench**, a unified benchmark for reasoning-based image editing assessment with broader evaluation dimension coverage and robust evaluation pipeline. We also design an automated multi-scenario data synthesis pipeline and construct **UniREdit-Data-100K**, a large-scale synthetic dataset with high-quality chain-of-thought (CoT) reasoning annotations. We fine-tune Bagel on this dataset and develop **UniREdit-Bagel**, demonstrating substantial improvements in both in-domain and out-of-distribution settings.
|
| 18 |
+
|
| 19 |
+
<img alt="image" src="https://github.com/Maplebb/UniREditBench/raw/main/docs/static/images/teaser.png" />
|
| 20 |
+
|
| 21 |
+
<img alt="image" src="https://github.com/Maplebb/UniREditBench/raw/main/docs/static/images/radar.png" />
|
| 22 |
+
|
| 23 |
+
### ✨ Highlights:
|
| 24 |
+
|
| 25 |
+
- **Broader Scenario and Reasoning Dimension Coverage**: It contains 2,700 high-quality samples organized into 8 primary reasoning dimensions and 18 sub-categories, spanning both real-world and game-world image editing tasks.
|
| 26 |
+
- **Reliable Dual-Reference Evaluation**: For each sample assessment, we design both the textual reference and ground-truth (GT) image reference. This multi-modal reference enables vision-language model (VLM) evaluators to perform direct and fine-grained comparisons at both the textual and visual levels with the generated images, leading to more reliable evaluation.
|
| 27 |
+
|
| 28 |
+
<img alt="image" src="https://github.com/Maplebb/UniREditBench/raw/main/docs/static/images/motivation_tab.png" />
|
| 29 |
+
<img alt="image" src="https://github.com/Maplebb/UniREditBench/raw/main/docs/static/images/motivation_fig.png" />
|
| 30 |
+
|
| 31 |
+
<img alt="image" src="https://github.com/Maplebb/UniREditBench/raw/main/docs/static/images/testpoint_cases.png" />
|
| 32 |
+
|
| 33 |
+
## 🚀 Sample Usage (Inference)
|
| 34 |
+
|
| 35 |
+
Here's an example of how to run inference with UniREdit-Bagel. Please refer to the [GitHub repository](https://github.com/Maplebb/UniREditBench) for detailed environment setup and checkpoint preparation steps.
|
| 36 |
+
|
| 37 |
+
```bash
|
| 38 |
+
GPUS=8
|
| 39 |
+
model_path=./ckpt
|
| 40 |
+
input_path=./UniREditBench
|
| 41 |
+
output_path=./output_images
|
| 42 |
+
|
| 43 |
+
# Image Editing with Reasoning
|
| 44 |
+
torchrun \
|
| 45 |
+
--nnodes=1 \
|
| 46 |
+
--nproc_per_node=$GPUS \
|
| 47 |
+
gen_images_mp_uniredit.py \
|
| 48 |
+
--input_dir $input_path \
|
| 49 |
+
--output_dir $output_path \
|
| 50 |
+
--metadata_file ./UniREditBench/data.json \
|
| 51 |
+
--max_latent_size 64 \
|
| 52 |
+
--model-path $model_path \
|
| 53 |
+
--think
|
| 54 |
+
```
|
| 55 |
+
|
| 56 |
+
## 📧 Contact
|
| 57 |
+
|
| 58 |
+
If you have any comments or questions, please open a new issue on the [GitHub repository](https://github.com/Maplebb/UniREditBench) or feel free to contact [Feng Han](fhan25@m.fudan.edu.cn) and [Yibin Wang](https://codegoat24.github.io).
|
| 59 |
+
|
| 60 |
+
## ⭐ Citation
|
| 61 |
+
|
| 62 |
+
If you find UniREditBench or UniREdit-Bagel useful, please cite our paper:
|
| 63 |
+
|
| 64 |
+
```bibtex
|
| 65 |
+
@misc{han2025unireditbench,
|
| 66 |
+
title={UniREditBench: A Unified Reasoning-based Image Editing Benchmark},
|
| 67 |
+
author={Feng Han and Yibin Wang and Chenglin Li and Zheming Liang and Dianyi Wang and Yang Jiao and Zhipeng Wei and Chao Gong and Cheng Jin and Jingjing Chen and Jiaqi Wang},
|
| 68 |
+
year={2025},
|
| 69 |
+
eprint={2511.01295},
|
| 70 |
+
archivePrefix={arXiv},
|
| 71 |
+
primaryClass={cs.CV},
|
| 72 |
+
url={https://arxiv.org/abs/2511.01295},
|
| 73 |
+
}
|
| 74 |
+
```
|