clt_pretrain_data / README.md
deqing's picture
Upload sharded deduplicated dataset (arrow-only)
aca933c verified
# Qwen3-Inspired Pre-training Dataset
## Overview
This dataset is a curated mixture of high-quality text data designed for large language model pre-training, inspired by the Qwen3 methodology. The dataset includes both training and validation splits.
## Dataset Statistics
**Total Size:** 10.42 billion tokens
- **Training Split:** 9.89 billion tokens (94.9%)
- **Validation Split:** 0.53 billion tokens (5.1%)
### Data Sources (Combined)
- **dclm_baseline**: 5.06B tokens (48.56%) - 4,088,916 documents
- **the_stack**: 1.65B tokens (15.79%) - 383,490 documents
- **common_corpus**: 1.5B tokens (14.36%) - 381,841 documents
- **mini_pile**: 1.43B tokens (13.73%) - 999,858 documents
- **math_pile**: 0.79B tokens (7.55%) - 72,936 documents
### Training Split Statistics
- **dclm_baseline**: 4.81B tokens (48.61%) - 3,884,088 documents
- **the_stack**: 1.58B tokens (15.97%) - 363,502 documents
- **common_corpus**: 1.42B tokens (14.37%) - 361,913 documents
- **mini_pile**: 1.36B tokens (13.78%) - 949,859 documents
- **math_pile**: 0.72B tokens (7.26%) - 68,947 documents
### Validation Split Statistics
- **dclm_baseline**: 0.25B tokens (47.69%) - 204,828 documents
- **common_corpus**: 0.08B tokens (14.22%) - 19,928 documents
- **math_pile**: 0.07B tokens (12.89%) - 3,989 documents
- **mini_pile**: 0.07B tokens (12.86%) - 49,999 documents
- **the_stack**: 0.07B tokens (12.33%) - 19,988 documents
## Data Processing Pipeline
1. **Data Collection**: Sourced from multiple high-quality datasets
2. **Standardization**: All data transformed to consistent format with `text`, `info`, and `source_data` fields
3. **Train/Validation Split**: Created 95%/5% splits within each source dataset
4. **Exact Deduplication**: Removed identical documents within each split
5. **Near Deduplication**: Used MinHashLSH with Jaccard similarity threshold of 0.85
6. **Quality Filtering**: Applied content-based filtering during processing
7. **Shuffling**: Applied shuffling within each large shard for better data distribution
## Data Format
Each example contains:
- `text`: The main text content
- `info`: Metadata from the original dataset (as string)
- `source_data`: Source dataset identifier
## Splits
The dataset contains two splits:
- `train`: Training data (95% of each source dataset)
- `validation`: Validation data (5% of each source dataset)
## Tokenization
Token counts were computed using the Llama3 tokenizer (`meta-llama/Meta-Llama-3-8B`).
## Usage
```python
from datasets import load_dataset
# Load the entire dataset
dataset = load_dataset("bluelightai-dev/qwen_clt_pretrain_data")
# Load specific splits
train_dataset = load_dataset("bluelightai-dev/qwen_clt_pretrain_data", split="train")
val_dataset = load_dataset("bluelightai-dev/qwen_clt_pretrain_data", split="validation")
```
## Dataset Sources
The dataset combines data from the following sources:
- **DCLM Baseline**: High-quality web text from DataComp-LM
- **Common Corpus**: Multilingual web text corpus
- **The Stack**: Deduplicated source code
- **Mini Pile**: Academic and reference texts
- **Math Pile**: Mathematical content and reasoning datasets
## License
Please refer to the individual source dataset licenses. This mixture is provided for research purposes.
## Citation
If you use this dataset, please cite the original source datasets and this work.