krisbailey's picture
Upload Parquet files and updated README
92af24d verified
---
license: apache-2.0
task_categories:
- text-generation
language:
- en
tags:
- redpajama
- llm
- dataset-reproduction
- redpajama-10b
- redpajama-subset
- redpajama-weighted
- redpajama-sample
- natural-language-processing
size_categories:
- 1B<n<10B
pretty_name: RedPajama 10B Weighted Subset
---
# RedPajama-10B-Weighted
A **canonical 10 Billion token weighted subset** of the [RedPajama-Data-1T](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T) dataset.
## Dataset Description
This dataset is a faithful reproduction of the original RedPajama-Data-1T distribution, scaled down to exactly **10 Billion tokens**. It is designed to preserve the **exact domain ratios** of the original dataset (excluding the defunct 'Books' subset). This allows researchers and developers to prototype, debug, and test on a representative slice of the data without needing to download or process the full 1 Terabyte dataset.
It serves as both a standalone dataset for medium-scale experiments and the parent source for smaller slices (like the [1B subset](https://huggingface.co/datasets/krisbailey/RedPajama-1B-Weighted)).
## Dataset Details
- **Total Tokens:** ~10,000,000,000 (10 Billion)
- **Source:** [togethercomputer/RedPajama-Data-1T](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T)
- **Language:** English
- **Format:** Apache Parquet
- **Producer:** Kris Bailey
## Motivation
The original RedPajama dataset is a standard for open-source LLM training, but its size (1TB+) makes it unwieldy for quick iteration, debugging, or educational purposes. Randomly sampling without care can destroy the delicate balance of data sources (CommonCrawl vs. C4 vs. GitHub).
This **RedPajama 10B subset** solves this by using a **weighted interleaving strategy** that strictly adheres to the original mixing ratios. It ensures that even at a smaller scale, the data seen by the model is distributionally equivalent to the full run.
## Dataset Creation Process
The creation process involved a precise streaming and interleaving pipeline:
### 1. Source Streaming
We streamed the data directly from `togethercomputer/RedPajama-Data-1T` to avoid local storage bottlenecks.
### 2. Weighted Interleaving
We defined the target probabilities based on the original token counts:
- **CommonCrawl:** 74.16%
- **C4:** 14.78%
- **GitHub:** 4.98%
- **ArXiv:** 2.36%
- **Wikipedia:** 2.03%
- **StackExchange:** 1.69%
An interleaving algorithm sampled from these streams according to these probabilities to construct a single, unified stream.
### 3. Buffer Shuffling
To avoid burstiness (e.g., seeing 1000 Wikipedia articles in a row), we implemented a **buffer shuffle** with a size of 10,000 documents. This ensures a healthy mixture of domains throughout the dataset.
### 4. Verification
The process ran until exactly 10 Billion tokens were collected. We verified that the final composition matches the target weights.
## Composition
| Subset | Weight | Approx. Tokens |
| :--- | :--- | :--- |
| **CommonCrawl** | 74.16% | ~7.42 B |
| **C4** | 14.78% | ~1.48 B |
| **GitHub** | 4.98% | ~0.50 B |
| **ArXiv** | 2.36% | ~0.24 B |
| **Wikipedia** | 2.03% | ~0.20 B |
| **StackExchange** | 1.69% | ~0.17 B |
## Usage
```python
from datasets import load_dataset
# Load the 10B weighted subset
ds = load_dataset("krisbailey/RedPajama-10B-Weighted", split="train")
print(ds)
```
## Citation
If you use this dataset, please cite the original RedPajama work:
```bibtex
@software{together2023redpajama,
author = {Together Computer},
title = {RedPajama: An Open Source Recipe to Reproduce LLaMA training dataset},
month = April,
year = 2023,
url = {https://github.com/togethercomputer/RedPajama-Data}
}
```