Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
The dataset viewer is not available for this split.
Error code:   UnexpectedError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Dataset Card for Mini Project Gutenberg (Cleaned English Subset, Tokenized) Dataset

This dataset is a mini subset of the dataset nikolina-p/gutenberg_flat, created for learning, testing streaming datasets, DDP training, and quick experimentation.

It is made from the first 24 books. The text is tokenized using OpenAI's tiktoken tokenizer. Its structure is adapted for training of autoregressive models in distributed environment: each split contains 8 shards, all shards within a split have the same number of tokens, and each row consists of 16×1,024 + 1 tokens.

Total number of tokens: 2.359.440

  • train split: 2.097.280,
  • validation split: 262.160

Usage

from datasets import load_dataset
ds = load_dataset("nikolina-p/mini_gutenberg_flat", split="train", streaming=True)
print(next(iter(ds)))
Downloads last month
2