Safetensors
File size: 2,414 Bytes
b6b4c3c
8c682fa
b6b4c3c
8c682fa
 
b6b4c3c
 
8c682fa
 
 
 
 
b6b4c3c
 
3dc7ee3
b6b4c3c
8c682fa
b6b4c3c
8c682fa
b6b4c3c
8c682fa
b6b4c3c
8c682fa
 
 
 
b6b4c3c
 
 
 
 
 
 
 
8c682fa
 
b6b4c3c
8c682fa
b6b4c3c
8c682fa
b6b4c3c
 
 
8c682fa
b6b4c3c
8c682fa
 
 
 
 
 
 
 
 
 
 
b6b4c3c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
base_model: Qwen/Qwen3-VL-8B-Instruct
license: mit
library_name: peft
pipeline_tag: image-text-to-text
---

# PVC-Judge: Pairwise Visual Consistency Judge

PVC-Judge is a state-of-the-art 8B assessment model for evaluating image editing models in visual consistency. It is a pairwise preference model designed to capture the preservation of identity, structure, and semantic coherence between edited and original images.

The model was introduced in the paper [GEditBench v2: A Human-Aligned Benchmark for General Image Editing](https://arxiv.org/abs/2603.28547) and is implemented as a LoRA adapter for [Qwen/Qwen3-VL-8B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-8B-Instruct).

<p align="center">
  <a href="https://arxiv.org/abs/2603.28547"><img src="https://img.shields.io/badge/Paper-arXiv%3A2603.28547-b31b1b?logo=arxiv&logoColor=red"></a>
  <a href="https://zhangqijiang07.github.io/gedit2_web/"><img src="https://img.shields.io/badge/%F0%9F%8C%90%20Project%20Page-Website-8A2BE2"></a>
  <a href="https://github.com/ZhangqiJiang07/GEditBench_v2"><img src="https://img.shields.io/badge/GitHub-Code-black?logo=github"></a>
  <a href="https://huggingface.co/datasets/GEditBench-v2/GEditBench-v2"><img src="https://img.shields.io/badge/%F0%9F%A4%97%20HF-GEditBench v2-blue"></a>
</p>

## 🚀 Quick Start

To use PVC-Judge, you typically need to merge the LoRA weights with the base model.

### 1. Merge LoRA weights
This step requires `torch`, `peft`, and `transformers`.

```bash
python ./scripts/merge_lora.py \
  --base-model-path /path/to/Qwen3/VL/8B/Instruct \
  --lora-weights-path /path/to/LoRA/Weights \
  --model-save-dir /path/to/save/PVC/Judge/model
```

### 2. Deployment or Local Inference
You can serve the merged model via vLLM or run local evaluation as described in the [official repository](https://github.com/ZhangqiJiang07/GEditBench_v2).

**Local Inference:**
```bash
# Setup environment
conda env create -f environments/pvc_judge.yml
conda activate pvc_judge

# Run evaluation
bash ./scripts/local_eval.sh vc_reward
```

## Citation

```bibtex
@article{jiang2026geditbenchv2,
  title={GEditBench v2: A Human-Aligned Benchmark for General Image Editing},
  author={Zhangqi Jiang and Zheng Sun and Xianfang Zeng and Yufeng Yang and Xuanyang Zhang and Yongliang Wu and Wei Cheng and Gang Yu and Xu Yang and Bihan Wen},
  journal={arXiv preprint arXiv:2603.28547},
  year={2026}
}
```