wikipedia / README.md
eminorhan's picture
Update README.md
5b033d7 verified
metadata
language:
  - en
dataset_info:
  - config_name: 100M_1
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 645762137.0362595
        num_examples: 225498
      - name: validation
        num_bytes: 2863715.5852214186
        num_examples: 1000
    download_size: 389527538
    dataset_size: 648625852.621481
  - config_name: 100M_2
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 646074282.0350486
        num_examples: 225607
      - name: validation
        num_bytes: 2863715.5852214186
        num_examples: 1000
    download_size: 389381161
    dataset_size: 648937997.62027
  - config_name: 100M_3
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 650730683.5766187
        num_examples: 227233
      - name: validation
        num_bytes: 2863715.5852214186
        num_examples: 1000
    download_size: 390292303
    dataset_size: 653594399.1618401
  - config_name: 10M_1
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 64820202.27148681
        num_examples: 22635
      - name: validation
        num_bytes: 2863715.5852214186
        num_examples: 1000
    download_size: 40370445
    dataset_size: 67683917.85670823
  - config_name: 10M_2
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 64236004.292101644
        num_examples: 22431
      - name: validation
        num_bytes: 2863715.5852214186
        num_examples: 1000
    download_size: 40412205
    dataset_size: 67099719.87732306
  - config_name: 10M_3
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 63250886.13078547
        num_examples: 22087
      - name: validation
        num_bytes: 2863715.5852214186
        num_examples: 1000
    download_size: 40514801
    dataset_size: 66114601.71600689
  - config_name: all
    features:
      - name: id
        dtype: string
      - name: url
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: train
        num_bytes: 18350156819
        num_examples: 6407814
    download_size: 10723740674
    dataset_size: 18350156819
configs:
  - config_name: 100M_1
    data_files:
      - split: train
        path: 100M_1/train-*
      - split: validation
        path: 100M_1/validation-*
  - config_name: 100M_2
    data_files:
      - split: train
        path: 100M_2/train-*
      - split: validation
        path: 100M_2/validation-*
  - config_name: 100M_3
    data_files:
      - split: train
        path: 100M_3/train-*
      - split: validation
        path: 100M_3/validation-*
  - config_name: 10M_1
    data_files:
      - split: train
        path: 10M_1/train-*
      - split: validation
        path: 10M_1/validation-*
  - config_name: 10M_2
    data_files:
      - split: train
        path: 10M_2/train-*
      - split: validation
        path: 10M_2/validation-*
  - config_name: 10M_3
    data_files:
      - split: train
        path: 10M_3/train-*
      - split: validation
        path: 10M_3/validation-*
  - config_name: all
    data_files:
      - split: train
        path: all/train-*

This repository contains random subsets of the English wikipedia obtained from "wikimedia/wikipedia" ("20231101.en"). It includes two random subsets of the English wikipedia, one containing roughly 10M words total (23k articles), the other containing roughly 100M words total (228K articles). These data are intended to be used for the BabyLM challenge. For convenience, the repository also includes the full English wikipedia containing roughly 2.8B words total (6.4M articles).

You can load these datasets as follows:

from datasets import load_dataset

ds_10M = load_dataset("eminorhan/wikipedia", "10M")  # 10M word subset

ds_100M = load_dataset("eminorhan/wikipedia", "100M")  # 100M word subset

ds_all = load_dataset("eminorhan/wikipedia", "all")  # the full data (2.8B words)

Both subsets come with train/validation splits, whereas the full data only has a train split. We applied lightweight preprocessing to the article texts using this script, which mainly strips away some sections of the articles like "References", "See also", etc.