| | --- |
| | size_categories: |
| | - n<1K |
| | --- |
| | 10K slice of OpenWebText - An open-source replication of the WebText dataset from OpenAI. |
| |
|
| | This is a small subset representing the first 10K records from the original dataset - created for testing. |
| |
|
| | The full 8M-record dataset is [here](https://huggingface.co/datasets/openwebtext). |
| |
|
| | ``` |
| | $ python -c "from datasets import load_dataset; ds=load_dataset('stas/openwebtext-10k'); print(ds)" |
| | DatasetDict({ |
| | train: Dataset({ |
| | features: ['text'], |
| | num_rows: 10000 |
| | }) |
| | }) |
| | ``` |
| |
|
| | * Records: 10,000 |
| | * compressed size: ~15MB |
| | * uncompressed size: 50MB |
| |
|
| | To convert to jsonlines: |
| |
|
| | ``` |
| | from datasets import load_dataset |
| | dataset_name = "stas/openwebtext-10k" |
| | name = dataset_name.split('/')[-1] |
| | ds = load_dataset(dataset_name, split='train') |
| | ds.to_json(f"{name}.jsonl", orient="records", lines=True) |
| | ``` |
| |
|
| | To see how this subset was created, here is the [instructions file](https://huggingface.co/datasets/stas/openwebtext-10k/blob/main/process.txt). |