mkd-hossain's picture
Upload README.md with huggingface_hub
0c9fe7e verified
metadata
license: other
task_categories:
  - text-generation
language:
  - ko
  - en
tags:
  - keural
  - moe
  - pretraining
  - binary-dataset
  - korean
  - multilingual
  - 69.5B-tokens
size_categories:
  - 10B<n<100B

Keural Stage 1 — 69.5 Billion Tokens Binary Dataset

Pre-tokenized binary dataset for the Keural MoE Foundation Model. Ready to use directly for training — no tokenization step needed.

Dataset Overview

Property Value
Total Tokens 69.5 billion (69,496,921,399)
Format Binary (.bin) + Index (.idx) + Metadata (.meta)
Total Sequences 15,761,448
Sequence Length 4,096 tokens
Shards 158 shards
Archive Size ~242GB (binary_69B_tokens.tar)
Tokenizer Keural SentencePiece Unigram, vocab=131,072
Last Updated 2026-04-03

Data Sources

Source Language Tokens (approx)
FineWeb English ~20B
WanJuan Korean Korean ~5B
Korean WebText Korean ~4B
ArXiv English Science ~4B
CC100 Korean Korean ~3B
PubMed English Medical ~3B
The Stack v1 Code ~8B
Wikipedia Korean Korean ~1B
PG19 Literature English ~1B
Other sources Mixed ~20.5B

Archive Contents

The tar file contains a binary/ folder with:

  • 158 .bin files: Pre-tokenized binary data (keural_000.bin to keural_157.bin)
  • 158 .idx files: Index files for fast random access
  • 158 .meta files: Metadata JSON for each shard
  • build_stats.json: Complete build statistics

Binary Format Specification

File: keural_NNN.bin
─────────────────────────────────────────
HEADER (36 bytes):
  [0:8]   magic   = b"KEURAL\x00\x00"   (8 bytes)
  [8:12]  version = 1                    (uint32 LE)
  [12:20] num_seq                        (uint64 LE)
  [20:28] seq_len = 4096                 (uint64 LE)
  [28:36] padding = 0                    (uint64 LE)

BODY:
  num_seq × 4096 × 4 bytes (uint32 LE tokens)

File: keural_NNN.idx
─────────────────────────────────────────
  [0:4]  num_seq  (uint32)
  [4:8]  seq_len  (uint32)
  per sequence: 8-byte offset + 4-byte length

File: keural_NNN.meta (JSON)
─────────────────────────────────────────
  {"num_sequences": N, "seq_length": 4096, "source": "keural_NNN"}

How to Extract

# Download the tar file, then extract:
tar -xf binary_69B_tokens.tar

# This creates: binary/ directory with 158 shards

How to Use

import struct, mmap, torch

HEADER_FMT  = "<8sIQQQ"
HEADER_SIZE = struct.calcsize(HEADER_FMT)  # 36 bytes

with open("binary/keural_001.bin", "rb") as f:
    raw = f.read(HEADER_SIZE)
    magic, ver, num_seqs, seq_len, _ = struct.unpack(HEADER_FMT, raw)
    print(f"Sequences: {num_seqs}, Length: {seq_len}")

# Or use directly with training scripts:
# torchrun --nproc_per_node=2 train_keural_v2.py --data_dir ./binary

Build Statistics

{
  "documents_processed": 553,711,744,
  "tokens_processed": 69,496,921,399,
  "sequences_written": 15,761,448,
  "padding_added": 3,143,055,066,
  "shards_created": 158,
  "sequence_utilization": "95.13%"
}

Related Resources

Author

Md Najmul Hossain / MKD CO., LTD.
Keural Foundation Model — Stage 1 pretraining dataset, 2026