| dataset_info: | |
| features: | |
| - name: text | |
| dtype: string | |
| splits: | |
| - name: train | |
| num_bytes: 5485226321 | |
| num_examples: 1000000 | |
| download_size: 3353329992 | |
| dataset_size: 5485226321 | |
| configs: | |
| - config_name: default | |
| data_files: | |
| - split: train | |
| path: data/train-* | |
| license: odc-by | |
| task_categories: | |
| - text-generation | |
| - fill-mask | |
| - feature-extraction | |
| language: | |
| - en | |
| size_categories: | |
| - 100K<n<1M | |
| # fineweb "longish" 1M | |
| 1m samples w/ random seed w.r.t. previous samples. | |
| - min 512 GPT-4 tiktoken tokens | |
| - max 8192 GPT-4 tiktoken tokens | |
| BEE-spoke-data/claude-tokenizer token count: | |
| ``` | |
| token_count | |
| count 1000000.000000 | |
| mean 1218.231641 | |
| std 935.733312 | |
| min 139.000000 | |
| 25% 683.000000 | |
| 50% 905.000000 | |
| 75% 1350.000000 | |
| max 9550.000000 | |
| ``` | |
| - Total count: 1218.23 M tokens |