Haon-Chen commited on
Commit
89b0e5a
Β·
1 Parent(s): 4361421

update README

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -17,7 +17,7 @@ pipeline_tag: zero-shot-image-classification
17
  This repo presents the `MoCa-Qwen25VL` series of **multimodal embedding models**.
18
  The model is trained based on [Qwen2.5-3B-VL-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-VL-Instruct).
19
 
20
- [🏠 Homepage](https://haon-chen.github.io/MoCa/) | [πŸ’» Code](https://github.com/haon-chen/MoCa) | [πŸ€– MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [πŸ€– MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [πŸ“š Datasets](https://huggingface.co/moca-embed/datasets) | [πŸ“„ Paper]()
21
 
22
  **Highlights**
23
  - SOTA performance on MMEB (General Multimodal) and surpassing many strong baselines on ViDoRe-v2 (Document Retrieval).
@@ -123,10 +123,10 @@ print(string, '=', compute_similarity(qry_output, tgt_output))
123
  ## Citation
124
  If you use this model in your research, please cite the associated paper.
125
  ```bibtex
126
- @article{xxx,
127
  title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
128
  author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
129
- journal={arXiv preprint arXiv:250xxxx},
130
  year={2025}
131
  }
132
  ```
 
17
  This repo presents the `MoCa-Qwen25VL` series of **multimodal embedding models**.
18
  The model is trained based on [Qwen2.5-3B-VL-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-VL-Instruct).
19
 
20
+ [🏠 Homepage](https://haon-chen.github.io/MoCa/) | [πŸ’» Code](https://github.com/haon-chen/MoCa) | [πŸ€– MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [πŸ€– MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [πŸ“š Datasets](https://huggingface.co/moca-embed/datasets) | [πŸ“„ Paper](https://arxiv.org/abs/2506.23115)
21
 
22
  **Highlights**
23
  - SOTA performance on MMEB (General Multimodal) and surpassing many strong baselines on ViDoRe-v2 (Document Retrieval).
 
123
  ## Citation
124
  If you use this model in your research, please cite the associated paper.
125
  ```bibtex
126
+ @article{chen2025moca,
127
  title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
128
  author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
129
+ journal={arXiv preprint arXiv:2506.23115},
130
  year={2025}
131
  }
132
  ```