Files changed (1) hide show
  1. README.md +28 -0
README.md CHANGED
@@ -2,6 +2,7 @@
2
  license: mit
3
  datasets:
4
  - timbrooks/instructpix2pix-clip-filtered
 
5
  language:
6
  - en
7
  base_model:
@@ -9,3 +10,30 @@ base_model:
9
  pipeline_tag: image-to-image
10
  library_name: diffusers
11
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: mit
3
  datasets:
4
  - timbrooks/instructpix2pix-clip-filtered
5
+ - Aleksandar/Top-Bench-X
6
  language:
7
  - en
8
  base_model:
 
10
  pipeline_tag: image-to-image
11
  library_name: diffusers
12
  ---
13
+
14
+ # EditCLIP: Representation Learning for Image Editing
15
+ [![Paper](https://img.shields.io/badge/arXiv-2503.20318-b31b1b)](https://arxiv.org/abs/2503.20318)
16
+ [![Project Page](https://img.shields.io/badge/🌐-Project_Page-blue)](https://qianwangx.github.io/EditCLIP/)
17
+ [![GitHub](https://img.shields.io/badge/GitHub-Repository-black?logo=github)](https://github.com/QianWangX/EditCLIP)
18
+ [![ICCV 2025](https://img.shields.io/badge/πŸ“·-Published_at_ICCV_2025-blue)](https://iccv2025.thecvf.com/)
19
+
20
+ ## πŸ’‘ Abstract
21
+
22
+ We introduce EditCLIP, a novel representation-learning approach for image editing. Our method learns a unified representation of edits by jointly encoding an input image and its edited counterpart, effectively capturing their transformation. To evaluate its effectiveness, we employ EditCLIP to solve two tasks: exemplar-based image editing and automated edit evaluation. In exemplar-based image editing, we replace text-based instructions in InstructPix2Pix with EditCLIP embeddings computed from a reference exemplar image pair. Experiments demonstrate that our approach outperforms state-of-the-art methods while being more efficient and versatile. For automated evaluation, EditCLIP assesses image edits by measuring the similarity between the EditCLIP embedding of a given image pair and either a textual editing instruction or the EditCLIP embedding of another reference image pair. Experiments show that EditCLIP aligns more closely with human judgments than existing CLIP-based metrics, providing a reliable measure of edit quality and structural preservation.
23
+
24
+ ## πŸ“Š Benchmark
25
+ We evaluate EditCLIP using **Top-Bench-X**, a benchmark for image editing evaluation:
26
+ - **Dataset:** Top-Bench-X
27
+ - **Link:** https://huggingface.co/datasets/Aleksandar/Top-Bench-X
28
+
29
+
30
+ ## 🌟 Citation
31
+ ```bibtex
32
+ @inproceedings{wang2025editclip,
33
+ title={EditCLIP: Representation Learning for Image Editing},
34
+ author={Wang, Qian and Cveji{\'c}, Aleksandar and Eldesokey, Abdelrahman and Wonka, Peter},
35
+ booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
36
+ pages={15960--15970},
37
+ year={2025}
38
+ }
39
+ ```