Datasets:

Modalities:
Text
Formats:
parquet
Languages:
Hindi
ArXiv:
License:
SultanR commited on
Commit
bf7a22f
·
verified ·
1 Parent(s): f04b7ef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -77
README.md CHANGED
@@ -1,115 +1,86 @@
1
  ---
2
- license: cc-by-4.0
3
  language:
4
  - hi
5
- size_categories:
6
- - 10M<n<100M
7
  task_categories:
8
  - text-generation
9
- tags:
10
- - pretraining
11
- - hindi
12
- - deduplication
13
- - quality-filtered
14
  configs:
15
  - config_name: minhash_deduped
16
  data_files:
17
- - split: train
18
- path: "minhash_deduped/**/*.parquet"
19
  - config_name: quality_filtered
20
  data_files:
21
- - split: train
22
- path: "quality_filtered/**/*.parquet"
23
- - config_name: consensus
24
  data_files:
25
- - split: train
26
- path: "consensus/*.parquet"
 
27
  ---
28
 
29
- # HinMix: Hindi Pretraining Data Mix
30
 
31
- A high-quality Hindi pretraining dataset created by combining, filtering, and deduplicating multiple sources.
32
 
33
- ## Dataset Description
34
-
35
- This dataset contains Hindi text from multiple web crawl sources, processed through a quality filtering and MinHash deduplication pipeline.
36
-
37
- ### Sources
38
- - **C4** (mC4 Hindi subset)
39
- - **CulturaX** (Hindi)
40
- - **Fineweb-2** (hin_Deva)
41
- - **HPLT-2** (hin_Deva)
42
- - **Sangraha** (verified and unverified Hindi splits)
43
 
44
  ## Subsets
45
 
46
- ### 1. `minhash_deduped` (Recommended)
47
- MinHash-deduplicated data. Each source was deduplicated individually to remove near-duplicate documents.
48
-
49
- ```python
50
- from datasets import load_dataset
51
- ds = load_dataset("AdaMLLab/HinMix", "minhash_deduped")
52
- ```
53
 
54
- **Statistics:**
55
- - ~60M documents
56
- - 136GB compressed
57
 
58
- ### 2. `quality_filtered`
59
- Quality-filtered data before deduplication. Use this if you want to apply your own deduplication.
60
 
61
  ```python
62
  from datasets import load_dataset
63
- ds = load_dataset("AdaMLLab/HinMix", "quality_filtered")
64
- ```
65
-
66
- **Statistics:**
67
- - ~99M documents
68
- - 231GB compressed
69
 
70
- ### 3. `consensus`
71
- Documents that appear in 2+ sources (exact text match). These are high-confidence documents verified across multiple crawls.
72
-
73
- ```python
74
- from datasets import load_dataset
75
- ds = load_dataset("AdaMLLab/HinMix", "consensus")
76
  ```
77
 
78
- **Statistics:**
79
- - 1.92M documents
80
- - 3.7GB compressed
81
 
82
- **Schema:**
83
- - `text`: Document text
84
- - `id`: Primary document ID
85
- - `sources`: List of sources where document appears (e.g., `["c4", "culturax"]`)
86
- - `all_ids`: All document IDs from all sources
87
- - `metadata`: Additional metadata
88
 
89
- ## Quality Filtering
 
 
 
 
 
 
 
 
90
 
91
- Documents were filtered based on:
92
- - Language identification (Hindi/Devanagari script ratio)
93
- - Document length constraints
94
- - Line quality metrics
95
- - Repetition detection
96
- - Boilerplate/policy phrase removal
97
 
98
- Filter thresholds based on Fineweb-2 Hindi configuration.
 
 
99
 
100
  ## Citation
101
 
102
- If you use this dataset, please cite:
103
-
104
- ```bibtex
105
- @dataset{hinmix2024,
106
- title={HinMix: Hindi Pretraining Data Mix},
107
- author={AdaMLLab},
108
- year={2024},
109
- publisher={Hugging Face}
 
110
  }
111
  ```
112
 
113
  ## License
114
 
115
- This dataset is released under CC-BY-4.0. Individual source datasets may have their own licenses.
 
1
  ---
 
2
  language:
3
  - hi
4
+ license: other
 
5
  task_categories:
6
  - text-generation
7
+ arxiv: 2512.18834
 
 
 
 
8
  configs:
9
  - config_name: minhash_deduped
10
  data_files:
11
+ - split: train
12
+ path: minhash_deduped/**/*.parquet
13
  - config_name: quality_filtered
14
  data_files:
15
+ - split: train
16
+ path: quality_filtered/**/*.parquet
17
+ - config_name: matched
18
  data_files:
19
+ - split: train
20
+ path: consensus/*.parquet
21
+ default: minhash_deduped
22
  ---
23
 
24
+ <img src="https://huggingface.co/datasets/AdaMLLab/HinMix/resolve/main/finetasks_hindi_main_results.png" width="900" alt="Finetasks benchmark scores, showing HinMix-MinHash as SOTA.">
25
 
26
+ HinMix ([https://arxiv.org/abs/2512.18834](https://arxiv.org/abs/2512.18834)) is a Hindi pretraining corpus containing 76 billion tokens across 60 million documents (in the minhash subset). Rather than scraping the web again, HinMix combines six publicly available Hindi datasets, applies Hindi-specific quality filtering, and performs cross-dataset deduplication.
27
 
28
+ We train a 1.4B parameter language model through nanotron on 30 billion tokens to show that HinMix outperforms the previous state-of-the-art, [CulturaX Hindi](https://huggingface.co/datasets/uonlp/CulturaX) (see [Appendix A9 in the Fineweb-2 paper](https://arxiv.org/pdf/2506.20920)). The `minhash_deduped` subset achieves an 11.6% relative improvement, while the `matched` subset achieves an 8.1% relative improvement.
 
 
 
 
 
 
 
 
 
29
 
30
  ## Subsets
31
 
32
+ | Subset | Documents | Tokens | Description |
33
+ |--------|-----------|--------|-------------|
34
+ | `quality_filtered` | 99.6M | 130.3B | Quality-filtered data before deduplication |
35
+ | `minhash_deduped` | 59.6M | 76.2B | Document-level MinHash deduplication |
36
+ | `matched` | 19.8M | 27.1B | Documents appearing in 2+ source datasets |
 
 
37
 
38
+ The matched subset uses cross-dataset agreement as a signal for quality.
 
 
39
 
40
+ ## Usage
 
41
 
42
  ```python
43
  from datasets import load_dataset
 
 
 
 
 
 
44
 
45
+ ds = load_dataset("AdaMLLab/HinMix", "minhash_deduped")
46
+ ds = load_dataset("AdaMLLab/HinMix", "quality_filtered")
47
+ ds = load_dataset("AdaMLLab/HinMix", "matched")
 
 
 
48
  ```
49
 
50
+ ## Sources
 
 
51
 
52
+ Tokens were counted using `meta-llama/Llama-3.2-3B`'s tokenizer.
 
 
 
 
 
53
 
54
+ | Source | Tokens (MinHash) | Documents (MinHash) |
55
+ |--------|------------------|---------------------|
56
+ | FineWeb-2 | 20.0B | 17.1M |
57
+ | CulturaX | 16.6B | 11.5M |
58
+ | Sangraha (unverified) | 11.5B | 8.9M |
59
+ | HPLT 2.0 | 10.2B | 6.7M |
60
+ | Sangraha (verified) | 10.1B | 9.1M |
61
+ | C4 | 7.7B | 6.3M |
62
+ | **Total** | **76.2B** | **59.6M** |
63
 
64
+ ## Pipeline
 
 
 
 
 
65
 
66
+ 1. Quality filtering with Hindi-specific thresholds (Devanagari script ratio, repetition patterns, language identification)
67
+ 2. Document-level MinHash deduplication (5-gram shingles, 14 bands, 8 hashes per band, similarity threshold 0.8)
68
+ 3. Cross-source matching to identify documents appearing in 2+ independent sources
69
 
70
  ## Citation
71
 
72
+ ```bib
73
+ @misc{alrashed2025mixminmatch,
74
+ title={Mix, MinHash, and Match: Cross-Source Agreement for Multilingual Pretraining Datasets},
75
+ author={Sultan Alrashed and Francesco Orabona},
76
+ year={2025},
77
+ eprint={2512.18834v2},
78
+ archivePrefix={arXiv},
79
+ primaryClass={cs.CL},
80
+ url={https://arxiv.org/abs/2512.18834v2},
81
  }
82
  ```
83
 
84
  ## License
85
 
86
+ See individual source dataset licenses.