update README
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ pipeline_tag: zero-shot-image-classification
|
|
| 16 |
This repo presents the `MoCa-Qwen25VL` series of **multimodal embedding models**.
|
| 17 |
The model is trained based on [Qwen2.5-7B-VL-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-VL-Instruct).
|
| 18 |
|
| 19 |
-
[π Homepage](https://haon-chen.github.io/MoCa/) | [π» Code](https://github.com/haon-chen/MoCa) | [π€ MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [π€ MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [π Datasets](https://huggingface.co/moca-embed/datasets) | [π Paper]()
|
| 20 |
|
| 21 |
**Highlights**
|
| 22 |
- SOTA performance on MMEB (General Multimodal) and surpassing many strong baselines on ViDoRe-v2 (Document Retrieval).
|
|
@@ -121,10 +121,10 @@ print(string, '=', compute_similarity(qry_output, tgt_output))
|
|
| 121 |
## Citation
|
| 122 |
If you use this model in your research, please cite the associated paper.
|
| 123 |
```bibtex
|
| 124 |
-
@article{
|
| 125 |
title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
|
| 126 |
author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
|
| 127 |
-
journal={arXiv preprint arXiv:
|
| 128 |
year={2025}
|
| 129 |
}
|
| 130 |
```
|
|
|
|
| 16 |
This repo presents the `MoCa-Qwen25VL` series of **multimodal embedding models**.
|
| 17 |
The model is trained based on [Qwen2.5-7B-VL-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-VL-Instruct).
|
| 18 |
|
| 19 |
+
[π Homepage](https://haon-chen.github.io/MoCa/) | [π» Code](https://github.com/haon-chen/MoCa) | [π€ MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [π€ MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [π Datasets](https://huggingface.co/moca-embed/datasets) | [π Paper](https://arxiv.org/abs/2506.23115)
|
| 20 |
|
| 21 |
**Highlights**
|
| 22 |
- SOTA performance on MMEB (General Multimodal) and surpassing many strong baselines on ViDoRe-v2 (Document Retrieval).
|
|
|
|
| 121 |
## Citation
|
| 122 |
If you use this model in your research, please cite the associated paper.
|
| 123 |
```bibtex
|
| 124 |
+
@article{chen2025moca,
|
| 125 |
title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
|
| 126 |
author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
|
| 127 |
+
journal={arXiv preprint arXiv:2506.23115},
|
| 128 |
year={2025}
|
| 129 |
}
|
| 130 |
```
|