Datasets:
Tasks:
Text Classification
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
10K - 100K
License:
| configs: | |
| - config_name: default | |
| data_files: | |
| - split: train | |
| path: data/train-* | |
| - split: test | |
| path: data/test-* | |
| - split: validation | |
| path: data/validation-* | |
| dataset_info: | |
| features: | |
| - name: word | |
| dtype: string | |
| - name: language | |
| dtype: string | |
| - name: input_ids | |
| sequence: int32 | |
| - name: attention_mask | |
| sequence: int8 | |
| - name: special_tokens_mask | |
| sequence: int8 | |
| - name: tokens | |
| sequence: string | |
| splits: | |
| - name: train | |
| num_bytes: 5310458 | |
| num_examples: 37849 | |
| - name: test | |
| num_bytes: 1981786 | |
| num_examples: 14123 | |
| - name: validation | |
| num_bytes: 2614514 | |
| num_examples: 18643 | |
| download_size: 2205128 | |
| dataset_size: 9906758 | |
| license: mit | |
| task_categories: | |
| - text-classification | |
| language: | |
| - en | |
| # Dataset Card for "english_char_split" | |
| This is a dataset of English words which have been tokenised by character. | |
| It was originally used to train a RoBERTa model from scratch on the masked language modelling task where during training, characters were randomly masked. | |
| This was ultimately used in an anomaly detection task where the embeddings from the trained model were used to detect non-English words - see full example [here](https://github.com/datasig-ac-uk/signature_applications/tree/master/anomaly_detection_language_dataset). |