Dataset Viewer

The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.

Dataset Card for Tokenized OLMoE Mix

Dataset Summary

This dataset contains pre-tokenized training and evaluation data designed for training custom small-scale OLMoE (Mixture-of-Experts) models.

The data is sourced primarily from the official AI2 Dolma 1 and C4 datasets and was curated to run ablation studies and reproduction experiments related to the OLMoE Technical Paper (arXiv:2409.02060). It is provided in .npy format, having been pre-tokenized using the allenai/gpt-neox-olmo-dolma-v1_5 tokenizer.

Dataset Structure

The dataset currently totals 7.38 GB and is split into three parts for easier handling:

  • part-0-00000.npy (2.51 GB)
  • part-0-00001.npy (4.29 GB)
  • part-0-00002.npy (574 MB)

Data Sources & Composition

Our training mix consists of approximately 4.7 Billion tokens in total, built from the following sources:

1. Training Data (3.689B tokens)

  • Source: A Wikipedia subset from Dolma 1.
  • Original HF Dataset: allenai/OLMoE-mix-0924
  • Command used to fetch raw data:
wget -O data/wiki-001.json.gz "[https://huggingface.co/datasets/allenai/OLMoE-mix-0924/resolve/main/data/wiki/wiki-0001.json.gz?download=true](https://huggingface.co/datasets/allenai/OLMoE-mix-0924/resolve/main/data/wiki/wiki-0001.json.gz?download=true)"
Downloads last month
22

Paper for iliasslasri/tokenized-OLMoE-mix