| --- |
| license: odc-by |
| task_categories: |
| - text-generation |
| - text2text-generation |
| language: |
| - en |
| --- |
| # C4 English Tokenized Samples |
|
|
| This dataset contains tokenized English samples from the C4 (Colossal Clean Crawled Corpus) dataset for natural language processing (NLP) tasks. |
|
|
| The first 125 000 entries from the `en` split of [allenai/c4](https://huggingface.co/datasets/allenai/c4) |
| were tokenized using [spaCy](https://spacy.io/)'s `en_core_web_sm` model. Tokens joined with spaces. |
|
|
| ## Features |
|
|
| - `text`: Original text from C4 |
| - `tokenized`: The tokenized and space-joined text |
| - `num_tokens`: Number of tokens after tokenization |
| - `num_punct_tokens`: Number of punctuation tokens after tokenization |
|
|
| ## Example |
|
|
| ```json |
| { |
| "text": "ALDUS MANUTIUS AND HIS THESAURUS CORNUCOPIAE OF 1496.\nSyracuse (1958) . 7.5 x 4.25, cloth, 32 pp, a v.g. copy [...]", |
| "tokenized": "ALDUS MANUTIUS AND HIS THESAURUS CORNUCOPIAE OF 1496 . \n Syracuse ( 1958 ) . 7.5 x 4.25 , cloth , 32 pp , a v.g . copy [...]", |
| "num_tokens": 84, |
| "num_punct_tokens": 19 |
| } |
| ``` |
|
|
| ## Usage |
|
|
| This dataset can be useful for: |
| - Text classification tasks |
| - Language modeling |
| - Sentiment analysis |
| - Other NLP applications requiring tokenized English text |
|
|
| Researchers and developers can use this dataset to jumpstart their projects without the need for initial tokenization steps. |
|
|
| ## Licence |
|
|
| This dataset is licensed under the ODC-BY (Open Data Commons Attribution) licence. |