wikipedia / README.md
eminorhan's picture
Update README.md
5b033d7 verified
---
language:
- en
dataset_info:
- config_name: 100M_1
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 645762137.0362595
num_examples: 225498
- name: validation
num_bytes: 2863715.5852214186
num_examples: 1000
download_size: 389527538
dataset_size: 648625852.621481
- config_name: 100M_2
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 646074282.0350486
num_examples: 225607
- name: validation
num_bytes: 2863715.5852214186
num_examples: 1000
download_size: 389381161
dataset_size: 648937997.62027
- config_name: 100M_3
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 650730683.5766187
num_examples: 227233
- name: validation
num_bytes: 2863715.5852214186
num_examples: 1000
download_size: 390292303
dataset_size: 653594399.1618401
- config_name: 10M_1
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 64820202.27148681
num_examples: 22635
- name: validation
num_bytes: 2863715.5852214186
num_examples: 1000
download_size: 40370445
dataset_size: 67683917.85670823
- config_name: 10M_2
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 64236004.292101644
num_examples: 22431
- name: validation
num_bytes: 2863715.5852214186
num_examples: 1000
download_size: 40412205
dataset_size: 67099719.87732306
- config_name: 10M_3
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 63250886.13078547
num_examples: 22087
- name: validation
num_bytes: 2863715.5852214186
num_examples: 1000
download_size: 40514801
dataset_size: 66114601.71600689
- config_name: all
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 18350156819
num_examples: 6407814
download_size: 10723740674
dataset_size: 18350156819
configs:
- config_name: 100M_1
data_files:
- split: train
path: 100M_1/train-*
- split: validation
path: 100M_1/validation-*
- config_name: 100M_2
data_files:
- split: train
path: 100M_2/train-*
- split: validation
path: 100M_2/validation-*
- config_name: 100M_3
data_files:
- split: train
path: 100M_3/train-*
- split: validation
path: 100M_3/validation-*
- config_name: 10M_1
data_files:
- split: train
path: 10M_1/train-*
- split: validation
path: 10M_1/validation-*
- config_name: 10M_2
data_files:
- split: train
path: 10M_2/train-*
- split: validation
path: 10M_2/validation-*
- config_name: 10M_3
data_files:
- split: train
path: 10M_3/train-*
- split: validation
path: 10M_3/validation-*
- config_name: all
data_files:
- split: train
path: all/train-*
---
This repository contains random subsets of the English wikipedia obtained from
[`"wikimedia/wikipedia"`](https://huggingface.co/datasets/wikimedia/wikipedia) (`"20231101.en"`).
It includes two random subsets of the English wikipedia, one containing roughly 10M words total (23k articles), the other containing roughly 100M words total (228K articles).
These data are intended to be used for the BabyLM challenge. For convenience, the repository also includes the full English wikipedia containing roughly 2.8B words total
(6.4M articles).
You can load these datasets as follows:
```python
from datasets import load_dataset
ds_10M = load_dataset("eminorhan/wikipedia", "10M") # 10M word subset
ds_100M = load_dataset("eminorhan/wikipedia", "100M") # 100M word subset
ds_all = load_dataset("eminorhan/wikipedia", "all") # the full data (2.8B words)
```
Both subsets come with `train`/`validation` splits, whereas the full data only has a `train` split.
We applied lightweight preprocessing to the article texts using [this script](https://github.com/eminorhan/babylm/blob/master/create_random_wikipedia.py),
which mainly strips away some sections of the articles like "References", "See also", *etc.*