| | --- |
| | dataset_info: |
| | features: |
| | - name: english_text |
| | dtype: string |
| | - name: language |
| | dtype: string |
| | - name: translated_text |
| | dtype: string |
| | - name: split |
| | dtype: string |
| | splits: |
| | - name: train |
| | num_bytes: 36890855 |
| | num_examples: 198084 |
| | - name: test |
| | num_bytes: 4071501 |
| | num_examples: 22009 |
| | download_size: 21823665 |
| | dataset_size: 40962356 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: train |
| | path: data/train-* |
| | - split: test |
| | path: data/test-* |
| | --- |
| | |
| |
|
| | ## Loading and Splitting Dataset To Various Languages |
| |
|
| | In this example, I will show you how to load the dataset and split by language for your downstream task. |
| |
|
| | ```python |
| | >>> from datasets import load_dataset |
| | |
| | >>> # load dataset |
| | >>> dataset = load_dataset("mosesdaudu/translation_dataset") |
| | |
| | >>> dataset |
| | DatasetDict({ |
| | train: Dataset({ |
| | features: ['english_text', 'language', 'translated_text', 'split'], |
| | num_rows: 198084 |
| | }) |
| | test: Dataset({ |
| | features: ['english_text', 'language', 'translated_text', 'split'], |
| | num_rows: 22009 |
| | }) |
| | }) |
| | |
| | >>> # Filter Dataset To Pidgin Language Only |
| | >>> pidgin_dataset = dataset.filter(lambda example: example['language'] == 'pidgin') |
| | |
| | >>> pidgin_dataset |
| | DatasetDict({ |
| | train: Dataset({ |
| | features: ['english_text', 'language', 'translated_text', 'split'], |
| | num_rows: 22476 |
| | }) |
| | test: Dataset({ |
| | features: ['english_text', 'language', 'translated_text', 'split'], |
| | num_rows: 2497 |
| | }) |
| | }) |
| | ``` |