# T2I-CompBench: A Comprehensive Benchmark for Open-world Compositional Text-to-image Generation Kaiyi Huang1, Kaiyue Sun1, Enze Xie2, Zhenguo Li2, and Xihui Liu1. **1The University of Hong Kong, 2Huawei Noah’s Ark Lab** ## 🚩 **New Features/Updates** - βœ… Dec. 02, 2023. Release the inference code for generating images in metric evaluation. - βœ… Oct. 20, 2023. πŸ’₯ Evaluation metric adopted by 🧨 [**DALL-E 3**](https://cdn.openai.com/papers/dall-e-3.pdf) as the evaluation metric for compositionality. - βœ… Sep. 30, 2023. πŸ’₯ Evaluation metric adopted by 🧨 [**PixArt-Ξ±**](https://arxiv.org/pdf/2310.00426.pdf) as the evaluation metric for compositionality. - βœ… Sep. 22, 2023. πŸ’₯ Paper accepted to Neurips 2023. - βœ… Jul. 9, 2023. Release the dataset, training and evaluation code. - [ ] Human evaluation of image-score pairs ## **Installing the dependencies** Before running the scripts, make sure to install the library's training dependencies: **Important** We recommend using the **latest code** to ensure consistency with the results presented in the paper. To make sure you can successfully run the example scripts, execute the following steps in a new virtual environment. We use the **diffusers version** as **0.15.0.dev0** You can either install the development version from PyPI: ```bash pip install diffusers==0.15.0.dev0 ``` or install from the provided source: ```bash unzip diffusers.zip cd diffusers pip install . ``` Then cd in the example folder and run ```bash pip install -r requirements.txt ``` And initialize an [πŸ€—Accelerate](https://github.com/huggingface/accelerate/) environment with: ```bash accelerate config ``` ## **Finetuning** 1. LoRA finetuning Use LoRA finetuning method, please refer to the link for downloading "lora_diffusion" directory: ``` https://github.com/cloneofsimo/lora/tree/master ``` 2. Example usage ``` export project_dir=/T2I-CompBench cd $project_dir export train_data_dir="examples/samples/" export output_dir="examples/output/" export reward_root="examples/reward/" export dataset_root="examples/dataset/color.txt" export script=GORS_finetune/train_text_to_image.py accelerate launch --multi_gpu --mixed_precision=fp16 \ --num_processes=8 --num_machines=1 \ --dynamo_backend=no "${script}" \ --train_data_dir="${train_data_dir}" \ --output_dir="${output_dir}" \ --reward_root="${reward_root}" \ --dataset_root="${dataset_root}" ``` or run ``` cd T2I-CompBench bash GORS_finetune/train.sh ``` The image directory should be a directory containing the images, e.g., ``` examples/samples/ β”œβ”€β”€ a green bench and a blue bowl_000000.png β”œβ”€β”€ a green bench and a blue bowl_000001.png └──... ``` The reward directory should include a json file named "vqa_result.json", and the json file should be a dictionary that maps from `{"question_id", "answer"}`, e.g., ``` [{"question_id": 0, "answer": "0.7110"}, {"question_id": 1, "answer": "0.7110"}, ...] ``` The dataset should be placed in the directory "examples/dataset/". ## **Evaluation** 1. Install the requirements MiniGPT4 is based on the repository, please refer to the link for environment dependencies and weights: ``` https://github.com/Vision-CAIR/MiniGPT-4 ``` 2. Example usage For evaluation, the input images files are stored in the directory "examples/samples/", with the format the same as the training data. #### BLIP-VQA: ``` export project_dir="BLIPvqa_eval/" cd $project_dir out_dir="examples/" python BLIP_vqa.py --out_dir=$out_dir ``` or run ``` cd T2I-CompBench bash BLIPvqa_eval/test.sh ``` The output files are formatted as a json file named "vqa_result.json" in "examples/annotation_blip/" directory. #### UniDet: download weight and put under repo experts/expert_weights: ``` mkdir -p UniDet_eval/experts/expert_weights cd UniDet_eval/experts/expert_weights wget https://huggingface.co/shikunl/prismer/resolve/main/expert_weights/Unified_learned_OCIM_RS200_6x%2B2x.pth ``` ``` export project_dir=UniDet_eval cd $project_dir python determine_position_for_eval.py ``` To calculate prompts from the **"complex" category**, set the **"--complex" parameter to True**; otherwise, set it to False. The output files are formatted as a json file named "vqa_result.json" in "examples/labels/annotation_obj_detection" directory. #### CLIPScore: ``` outpath="examples/" python CLIPScore_eval/CLIP_similarity.py --outpath=${outpath} ``` or run ``` cd T2I-CompBench bash CLIPScore_eval/test.sh ``` To calculate prompts from the **"complex" category**, set the **"--complex" parameter to True**; otherwise, set it to False. The output files are formatted as a json file named "vqa_result.json" in "examples/annotation_clip" directory. #### 3-in-1: ``` export project_dir="3_in_1_eval/" cd $project_dir outpath="examples/" python "3_in_1.py" --outpath=${outpath} ``` The output files are formatted as a json file named "vqa_result.json" in "examples/annotation_3_in_1" directory. #### MiniGPT4-CoT: If the category to be evaluated is one of color, shape and texture: ``` export project_dir=Minigpt4_CoT_eval cd $project_dir category="color" img_file="examples/samples/" output_path="examples/" python mGPT_cot_attribute.py --category=${category} --img_file=${img_file} --output_path=${output_path} ``` If the category to be evaluated is one of spatial, non-spatial and complex: ``` export project_dir=MiniGPT4_CoT_eval/ cd $project_dir category="non-spatial" img_file="examples/samples/" output_path="examples" python mGPT_cot_general.py --category=${category} --img_file=${img_file} --output_path=${output_path} ``` The output files are formatted as a csv file named "mGPT_cot_output.csv" in output_path. ### Inference Run the inference.py to visualize the image. ``` export pretrained_model_path="checkpoint/color/lora_weight_e357_s124500.pt.pt" export prompt="A bathroom with green tile and a red shower curtain" python inference.py --pretrained_model_path "${pretrained_model_path}" --prompt "${prompt}" ``` **Generate images for metric calculation.** Run the inference_eval.py to generate images in the test set. As stated in the paper, 10 images are generated per prompt for **metric calculation**, and we use the fixed seed across all methods. You can specify the test set by changing the "from_file" parameter among {color_val.txt, shape_val.txt, texture_val.txt, spatial_val.txt, non_spatial_val.txt, complex_val.txt}. ``` export from_file="../examples/dataset/color_val.txt" python inference_eval.py --from_file "${from_file}" ``` ### Citation If you're using T2I-CompBench in your research or applications, please cite using this BibTeX: ```bibtex @article{huang2023t2icompbench, title={T2I-CompBench: A Comprehensive Benchmark for Open-world Compositional Text-to-image Generation}, author={Kaiyi Huang and Kaiyue Sun and Enze Xie and Zhenguo Li and Xihui Liu}, journal={arXiv preprint arXiv:2307.06350}, year={2023}, } ``` ### License This project is licensed under the MIT License. See the "License.txt" file for details.