Haon-Chen commited on
Commit
611a540
Β·
1 Parent(s): 27383e9

update README

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -16,7 +16,7 @@ pipeline_tag: zero-shot-image-classification
16
  This repo presents the `MoCa-Qwen25VL` series of **multimodal embedding models**.
17
  The model is trained based on [Qwen2.5-7B-VL-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-VL-Instruct).
18
 
19
- [🏠 Homepage](https://haon-chen.github.io/MoCa/) | [πŸ’» Code](https://github.com/haon-chen/MoCa) | [πŸ€– MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [πŸ€– MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [πŸ“š Datasets](https://huggingface.co/moca-embed/datasets) | [πŸ“„ Paper]()
20
 
21
  **Highlights**
22
  - SOTA performance on MMEB (General Multimodal) and surpassing many strong baselines on ViDoRe-v2 (Document Retrieval).
@@ -121,10 +121,10 @@ print(string, '=', compute_similarity(qry_output, tgt_output))
121
  ## Citation
122
  If you use this model in your research, please cite the associated paper.
123
  ```bibtex
124
- @article{xxx,
125
  title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
126
  author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
127
- journal={arXiv preprint arXiv:250xxxx},
128
  year={2025}
129
  }
130
  ```
 
16
  This repo presents the `MoCa-Qwen25VL` series of **multimodal embedding models**.
17
  The model is trained based on [Qwen2.5-7B-VL-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-VL-Instruct).
18
 
19
+ [🏠 Homepage](https://haon-chen.github.io/MoCa/) | [πŸ’» Code](https://github.com/haon-chen/MoCa) | [πŸ€– MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [πŸ€– MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [πŸ“š Datasets](https://huggingface.co/moca-embed/datasets) | [πŸ“„ Paper](https://arxiv.org/abs/2506.23115)
20
 
21
  **Highlights**
22
  - SOTA performance on MMEB (General Multimodal) and surpassing many strong baselines on ViDoRe-v2 (Document Retrieval).
 
121
  ## Citation
122
  If you use this model in your research, please cite the associated paper.
123
  ```bibtex
124
+ @article{chen2025moca,
125
  title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
126
  author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
127
+ journal={arXiv preprint arXiv:2506.23115},
128
  year={2025}
129
  }
130
  ```