Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
Dask
Haon-Chen commited on
Commit
0eeaec8
ยท
verified ยท
1 Parent(s): 3caf8b9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -4
README.md CHANGED
@@ -7,13 +7,55 @@ dataset_info:
7
  sequence: binary
8
  splits:
9
  - name: train
10
- num_bytes: 93686373572
11
- num_examples: 1448799
12
- download_size: 63925876676
13
- dataset_size: 93686373572
14
  configs:
15
  - config_name: default
16
  data_files:
17
  - split: train
18
  path: data/train-*
19
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  sequence: binary
8
  splits:
9
  - name: train
10
+ num_bytes: 94020556918
11
+ num_examples: 1465964
12
+ download_size: 73033984223
13
+ dataset_size: 94020556918
14
  configs:
15
  - config_name: default
16
  data_files:
17
  - split: train
18
  path: data/train-*
19
  ---
20
+ # MMEB train split used in MoCa Continual Pre-training
21
+
22
+ [๐Ÿ  Homepage](https://haon-chen.github.io/MoCa/) | [๐Ÿ’ป Code](https://github.com/haon-chen/MoCa) | [๐Ÿค– MoCa-Qwen25VL-7B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-7B) | [๐Ÿค– MoCa-Qwen25VL-3B](https://huggingface.co/moca-embed/MoCa-Qwen25VL-3B) | [๐Ÿ“š Datasets](https://huggingface.co/moca-embed/datasets) | [๐Ÿ“„ Paper](https://arxiv.org/abs/2506.23115)
23
+
24
+ ## Introduction
25
+
26
+ This is a interleaved multimodal pre-training dataset used in the modality-aware continual pre-training of MoCa models. It is adapted from the train split of [
27
+ MMEB](https://huggingface.co/datasets/TIGER-Lab/MMEB-train) by concatenating queries and positive documents.
28
+
29
+ The dataset consists of interleaved multimodal examples. text is a string containing text while images are image binaries that can be loaded with the following code snippet:
30
+
31
+ ```python
32
+ import PIL.Image
33
+ from io import BytesIO
34
+
35
+ image_bytes = example['images'][0]
36
+ image = PIL.Image.open(BytesIO(image_bytes))
37
+ ```
38
+
39
+
40
+ ## Citation
41
+ MoCa
42
+
43
+ ```bibtex
44
+ @article{chen2025moca,
45
+ title={MoCa: Modality-aware Continual Pre-training Makes Better Bidirectional Multimodal Embeddings},
46
+ author={Chen, Haonan and Liu, Hong and Luo, Yuping and Wang, Liang and Yang, Nan and Wei, Furu and Dou, Zhicheng},
47
+ journal={arXiv preprint arXiv:2506.23115},
48
+ year={2025}
49
+ }
50
+ ```
51
+
52
+ MMEB
53
+
54
+ ```bibtex
55
+ @article{jiang2024vlm2vec,
56
+ title={VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks},
57
+ author={Jiang, Ziyan and Meng, Rui and Yang, Xinyi and Yavuz, Semih and Zhou, Yingbo and Chen, Wenhu},
58
+ journal={arXiv preprint arXiv:2410.05160},
59
+ year={2024}
60
+ }
61
+ ```