|
|
--- |
|
|
tags: |
|
|
- clip |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
library_name: transformers |
|
|
pipeline_tag: zero-shot-image-classification |
|
|
--- |
|
|
# FG-CLIP: Fine-Grained Visual and Textual Alignment |
|
|
**[FG-CLIP: Fine-Grained Visual and Textual Alignment](https://arxiv.org/abs/2505.05071)** |
|
|
</br> |
|
|
Chunyu Xie*, Bin Wang*, Fanjing Kong, Jincheng Li, Dawei Liang, Gengshen Zhang, Dawei Leng†, Yuhui Yin(*Equal Contribution, ✝Corresponding Author) |
|
|
</br> |
|
|
[](https://arxiv.org/abs/2505.05071) |
|
|
[](https://icml.cc/Conferences/2025) |
|
|
[](https://github.com/360CVGroup/FG-CLIP) |
|
|
|
|
|
<p align="center"> |
|
|
<img src="https://huggingface.co/qihoo360/fg-clip-large/resolve/main/radar_chart_methods.png" width="500" height="440"/> |
|
|
</p> |
|
|
|
|
|
## Model Framework |
|
|
FG-CLIP’s training proceeds in two stages: the first stage leverages |
|
|
global-level caption-image pairs to achieve initial fine-grained alignment, while the second stage supplements these with additional |
|
|
region-level captions, including detailed region captions and positive/negative region descriptions to further refine the alignment. |
|
|
<p align="center"> |
|
|
<img src="https://huggingface.co/qihoo360/fg-clip-large/resolve/main/fgclip_strc.png" width=80%/> |
|
|
</p> |
|
|
|
|
|
## Quick Start 🤗 |
|
|
|
|
|
### Load Model |
|
|
```Shell |
|
|
import torch |
|
|
from PIL import Image |
|
|
from transformers import ( |
|
|
AutoImageProcessor, |
|
|
AutoTokenizer, |
|
|
AutoModelForCausalLM, |
|
|
) |
|
|
|
|
|
|
|
|
model_root = "qihoo360/fg-clip-base" |
|
|
image_size=224 |
|
|
model = AutoModelForCausalLM.from_pretrained(model_root,trust_remote_code=True).cuda() |
|
|
|
|
|
device = model.device |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_root) |
|
|
image_processor = AutoImageProcessor.from_pretrained(model_root) |
|
|
``` |
|
|
|
|
|
|
|
|
### Retrieval |
|
|
|
|
|
```Shell |
|
|
|
|
|
img_root = "FG-CLIP/use_imgs/cat_dfclor.jpg" |
|
|
image = Image.open(img_root).convert("RGB") |
|
|
image = image.resize((image_size,image_size)) |
|
|
|
|
|
image_input = image_processor.preprocess(image, return_tensors='pt')['pixel_values'].to(device) |
|
|
|
|
|
# NOTE Short captions: max_length=77 && walk_short_pos=True |
|
|
walk_short_pos = True |
|
|
captions=["a photo of a cat", "a photo of a dog"] |
|
|
caption_input = torch.tensor(tokenizer(captions, max_length=77, padding="max_length", truncation=True).input_ids, dtype=torch.long, device=device) |
|
|
|
|
|
# NOTE Long captions: max_length=248 && walk_short_pos=False |
|
|
# ...... |
|
|
|
|
|
with torch.no_grad(): |
|
|
image_feature = model.get_image_features(image_input) |
|
|
text_feature = model.get_text_features(caption_input,walk_short_pos=walk_short_pos) |
|
|
image_feature = image_feature / image_feature.norm(p=2, dim=-1, keepdim=True) |
|
|
text_feature = text_feature / text_feature.norm(p=2, dim=-1, keepdim=True) |
|
|
|
|
|
logits_per_image = image_feature @ text_feature.T |
|
|
logits_per_image = model.logit_scale.exp() * logits_per_image |
|
|
probs = logits_per_image.softmax(dim=1) |
|
|
print(probs) |
|
|
# [[9.9997e-01, 3.3485e-05]] |
|
|
``` |
|
|
|
|
|
### Dense feature effect display |
|
|
|
|
|
```Shell |
|
|
|
|
|
import math |
|
|
import matplotlib |
|
|
matplotlib.use('Agg') |
|
|
import matplotlib.pyplot as plt |
|
|
|
|
|
|
|
|
img_root = "FG-CLIP/use_imgs/cat_dfclor.jpg" |
|
|
image = Image.open(img_root).convert("RGB") |
|
|
image = image.resize((image_size,image_size)) |
|
|
|
|
|
image_input = image_processor.preprocess(image, return_tensors='pt')['pixel_values'].to(device) |
|
|
|
|
|
with torch.no_grad(): |
|
|
dense_image_feature = model.get_image_dense_features(image_input) |
|
|
captions = ["white cat"] |
|
|
caption_input = torch.tensor(tokenizer(captions, max_length=77, padding="max_length", truncation=True).input_ids, dtype=torch.long, device=device) |
|
|
text_feature = model.get_text_features(caption_input,walk_short_pos=True) |
|
|
text_feature = text_feature / text_feature.norm(p=2, dim=-1, keepdim=True) |
|
|
dense_image_feature = dense_image_feature / dense_image_feature.norm(p=2, dim=-1, keepdim=True) |
|
|
|
|
|
similarity = dense_image_feature.squeeze() @ text_feature.squeeze().T |
|
|
similarity = similarity.cpu().numpy() |
|
|
patch_size = int(math.sqrt(similarity.shape[0])) |
|
|
|
|
|
|
|
|
original_shape = (patch_size, patch_size) |
|
|
show_image = similarity.reshape(original_shape) |
|
|
|
|
|
|
|
|
plt.figure(figsize=(6, 6)) |
|
|
plt.imshow(show_image) |
|
|
plt.title('similarity Visualization') |
|
|
plt.axis('off') |
|
|
plt.savefig("FG-CLIP/use_imgs/FGCLIP_dfcolor_cat.png") |
|
|
|
|
|
``` |
|
|
<!-- /home/jovyan/wangbin-home-shcdt/image_text_match/FG-CLIP/use_imgs/FGCLIP_dfcolor_cat.png --> |
|
|
<p align="left"> |
|
|
<img src="https://huggingface.co/qihoo360/fg-clip-large/resolve/main/FGCLIP_dfcolor_cat.png" width=25%/> |
|
|
</p> |
|
|
|
|
|
## Citation |
|
|
If you find FG-CLIP useful for your research and applications, please cite using this BibTeX: |
|
|
|
|
|
``` |
|
|
@article{xie2025fgclip, |
|
|
title={FG-CLIP: Fine-Grained Visual and Textual Alignment}, |
|
|
author={Chunyu Xie and Bin Wang and Fanjing Kong and Jincheng Li and Dawei Liang and Gengshen Zhang and Dawei Leng and Yuhui Yin}, |
|
|
year={2025}, |
|
|
eprint={2505.05071}, |
|
|
archivePrefix={arXiv}, |
|
|
primaryClass={cs.CV}, |
|
|
url={https://arxiv.org/abs/2505.05071}, |
|
|
} |
|
|
``` |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
## License |
|
|
|
|
|
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses. |
|
|
The content of this project itself is licensed under the [Apache license 2.0](./LICENSE). |
|
|
|
|
|
|