|
|
--- |
|
|
license: unknown |
|
|
base_model: |
|
|
- black-forest-labs/FLUX.1-Fill-dev |
|
|
- black-forest-labs/FLUX.1-dev |
|
|
- black-forest-labs/FLUX.1-Redux-dev |
|
|
--- |
|
|
# [NeurIPS 2025] Domain-RAG: Retrieval-Guided Compositional Image Generation for Cross-Domain Few-Shot Object Detection |
|
|
|
|
|
[π₯ Paper (NeurIPS 2025)](https://arxiv.org/abs/2506.05872) | [π Project Page](https://yuli-cs.net/papers/domain-rag) | [π¦ Dataset Scripts](#dataset-preparation) | [π§ Model Zoo](#pretrained-models) | [π Quick Start](#quick-start) | [π₯ Video](#video) | [π Evaluation](#evaluation) | [π Contact](#contact) |
|
|
|
|
|
--- |
|
|
|
|
|
**Domain-RAG** is a novel retrieval-augmented generative framework designed for **Cross-Domain Few-Shot Object Detection (CD-FSOD)**. We leverage large-scale vision-language models (GroundingDINO), a curated COCO-style retrieval corpus, and Flux-based background generation to synthesize diverse, domain-aware training data that enhances FSOD generalization under domain shift. |
|
|
|
|
|
|
|
|
--- |
|
|
|
|
|
## β¨ Highlights |
|
|
|
|
|
- π **Retrieval-Augmented Generation**: retrieve semantically similar source images for novel-class prompts. |
|
|
- π¨ **Flux-Redux Integration**: compose diverse backgrounds with target foregrounds for domain-aligned generation. |
|
|
- π¦ **Support for Multiple Target Domains**: ArTAXOr, Clipart1k, DIOR, DeepFish, UODD, NEU-DET, and more. |
|
|
- π§ͺ **Strong Benchmarks**: surpasses GroundingDINO baseline in 1-shot and 5-shot CD-FSOD across 6 datasets. |
|
|
|
|
|
--- |
|
|
|
|
|
## π§ Installation |
|
|
|
|
|
```bash |
|
|
git clone https://github.com/LiYu0524/Domain-RAG.git |
|
|
cd Domain-RAG |
|
|
conda create -n domainrag python=3.10 |
|
|
conda activate domainrag |
|
|
pip install -r requirements.txt |
|
|
``` |
|
|
|
|
|
|
|
|
## Pretrained Models |
|
|
|
|
|
we will relase the fine-tuned grounding-dino model soon |
|
|
|
|
|
## Dataset Preparation |
|
|
|
|
|
You can prepare CDFSOD with [CDVITO](https://github.com/lovelyqian/CDFSOD-benchmark?tab=readme-ov-file) |
|
|
|
|
|
## Quick start |
|
|
|
|
|
You can refer to `./domainrag.sh` |
|
|
|
|
|
|
|
|
|
|
|
## Video |
|
|
|
|
|
Walkthrough video(Chinese version): [Watch here](https://www.bilibili.com/video/BV1YznKzkEEK/?spm_id_from=333.337.search-card.all.click&vd_source=23bede4ceb3dc1ea2ffc645933850555) |
|
|
|
|
|
## Contact |
|
|
|
|
|
For questions and collaboration, please contact: |
|
|
|
|
|
- Yu Li : `<liyu24@m.fudan.edu.cn>` |
|
|
|
|
|
|
|
|
## Citation |
|
|
|
|
|
If you find **Domain-RAG** useful in your research, please cite: |
|
|
|
|
|
```bibtex |
|
|
@inproceedings{li2025domainrag, |
|
|
author={Li, Yu and Qiu, Xingyu and Fu, Yuqian and Chen, Jie and Qian, Tianwen and Zheng, Xu and Paudel, Danda Pani and Fu, Yanwei and Huang, Xuanjing and Van Gool, Luc and others}, |
|
|
booktitle = {Advances in Neural Information Processing Systems}, |
|
|
title = {Domain-RAG: Retrieval-Guided Compositional Image Generation for Cross-Domain Few-Shot Object Detection}, |
|
|
year = {2025} |
|
|
} |
|
|
|
|
|
``` |
|
|
|
|
|
If you find **CD-Vito** useful in your research, please cite: |
|
|
```bibtex |
|
|
@inproceedings{fu2024cross, |
|
|
title={Cross-domain few-shot object detection via enhanced open-set object detector}, |
|
|
author={Fu, Yuqian and Wang, Yu and Pan, Yixuan and Huai, Lian and Qiu, Xingyu and Shangguan, Zeyu and Liu, Tong and Fu, Yanwei and Van Gool, Luc and Jiang, Xingqun}, |
|
|
booktitle={European Conference on Computer Vision}, |
|
|
pages={247--264}, |
|
|
year={2024}, |
|
|
organization={Springer} |
|
|
} |
|
|
``` |