| | --- |
| | dataset_info: |
| | features: |
| | - name: video_id |
| | dtype: string |
| | - name: video_link |
| | dtype: string |
| | - name: title |
| | dtype: string |
| | - name: text |
| | dtype: string |
| | - name: channel |
| | dtype: string |
| | - name: channel_id |
| | dtype: string |
| | - name: date |
| | dtype: string |
| | - name: license |
| | dtype: string |
| | - name: original_language |
| | dtype: string |
| | - name: language_id_method |
| | dtype: string |
| | - name: transcription_language |
| | dtype: string |
| | - name: word_count |
| | dtype: int64 |
| | - name: character_count |
| | dtype: int64 |
| | - name: source_language |
| | dtype: string |
| | splits: |
| | - name: train |
| | num_bytes: 298197594003 |
| | num_examples: 22684737 |
| | download_size: 162573072184 |
| | dataset_size: 298197594003 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: train |
| | path: data/train-* |
| | license: cc-by-4.0 |
| | task_categories: |
| | - text-generation |
| | tags: |
| | - conversational |
| | language: |
| | - en |
| | - fr |
| | - es |
| | - pt |
| | - de |
| | - ru |
| | - nl |
| | - tr |
| | - it |
| | pretty_name: YouTube Commons Re-upload |
| | --- |
| | |
| | ## YouTube Commons Re-upload |
| |
|
| | This is a re-upload of [PleIAs' YouTube Commons](https://huggingface.co/datasets/PleIAs/YouTube-Commons), a valuable open dataset: |
| | > YouTube-Commons is a collection of audio transcripts of 2,063,066 videos shared on YouTube under a CC BY 4.0 license. |
| | > |
| | > **Content** |
| | > |
| | > The collection comprises 22,709,724 original and automatically translated transcripts from 3,156,703 videos (721,136 individual channels). |
| |
|
| | Unfortunately, there are [problems](https://huggingface.co/datasets/PleIAs/YouTube-Commons/discussions/10) with loading YouTube Commons with Hugging Face Datasets. |
| | In order to alleviate those and to further process the dataset, I took the source parquet-files and reuploaded this fixed version to HuggingFace. |
| |
|
| | ## Code |
| | The code used for this reupload. It makes use of a git clone of the [PleIAs/YouTube-Commons](https://huggingface.co/datasets/PleIAs/YouTube-Commons) dataset. |
| |
|
| | ```python |
| | from pathlib import Path |
| | |
| | from datasets import load_dataset, Dataset |
| | from tqdm import tqdm |
| | |
| | columns = set('''video_link |
| | video_id |
| | title |
| | text |
| | channel |
| | channel_id |
| | date |
| | license |
| | original_language |
| | language_id_method |
| | transcription_language |
| | source_language |
| | word_count |
| | character_count'''.split('\n')) |
| | |
| | def generate(): |
| | for filepath in tqdm(sorted(Path('/Path/To/PleIAs/YouTube-Commons').rglob('*.parquet'))): |
| | print(filepath) |
| | dataset = load_dataset("parquet", |
| | data_files={'train': str(filepath)}) |
| | for row in dataset['train']: |
| | keys = set(row) |
| | # Some of the files are missing one of these two columns. |
| | # Setting them to None results in an Arrow error, so we use '' instead |
| | if 'language_id_method' not in keys: |
| | row['language_id_method'] = '' |
| | if 'source_language' not in keys: |
| | row['source_language'] = '' |
| | if '__index_level_0__' in keys: |
| | del row['__index_level_0__'] |
| | |
| | if not set(row) == columns: |
| | raise ValueError(f'Error in columns: {set(row)}') |
| | yield row |
| | |
| | youtube = Dataset.from_generator(generate) |
| | youtube.push_to_hub('Rijgersberg/YouTube-Commons') |
| | |
| | ``` |