Datasets:
ArXiv:
License:
| #!/usr/bin/python3 | |
| # -*- coding: utf-8 -*- | |
| import argparse | |
| import json | |
| import os | |
| import random | |
| import sys | |
| pwd = os.path.abspath(os.path.dirname(__file__)) | |
| sys.path.append(os.path.join(pwd, "../../../")) | |
| from datasets import load_dataset, DownloadMode | |
| from project_settings import project_path | |
| def get_args(): | |
| parser = argparse.ArgumentParser() | |
| parser.add_argument( | |
| "--dataset_cache_dir", | |
| default=(project_path / "hub_datasets").as_posix(), | |
| type=str | |
| ) | |
| parser.add_argument( | |
| "--train_subset", | |
| default="train.jsonl", | |
| type=str | |
| ) | |
| parser.add_argument( | |
| "--valid_subset", | |
| default="valid.jsonl", | |
| type=str | |
| ) | |
| args = parser.parse_args() | |
| return args | |
| s = """ | |
| | ar | arabic | 10000 | iwslt2017 | | |
| | bg | bulgarian | 10000 | xnli | | |
| | bn | bengali | 10000 | open_subtitles | | |
| | bs | bosnian | 10000 | open_subtitles | | |
| | cs | czech | 10000 | ecb | | |
| | da | danish | 10000 | open_subtitles | | |
| | de | german | 10000 | ecb | | |
| | el | modern greek | 10000 | ecb | | |
| | en | english | 10000 | ecb | | |
| | eo | esperanto | 10000 | tatoeba | | |
| | es | spanish | 10000 | tatoeba | | |
| | et | estonian | 10000 | emea | | |
| | fi | finnish | 10000 | ecb | | |
| | fo | faroese | 10000 | nordic_langid | | |
| | fr | french | 10000 | iwslt2017 | | |
| | ga | irish | 10000 | multi_para_crawl | | |
| | gl | galician | 3096 | tatoeba | | |
| | hi | hindi | 10000 | open_subtitles | | |
| | hi_en | hindi | 7180 | cmu_hinglish_dog | | |
| | hr | croatian | 10000 | hrenwac_para | | |
| | hu | hungarian | 3801 | europa_ecdc_tm; europa_eac_tm | | |
| | hy | armenian | 660 | open_subtitles | | |
| | id | indonesian | 10000 | id_panl_bppt | | |
| | is | icelandic | 2973 | europa_ecdc_tm; europa_eac_tm | | |
| | it | italian | 10000 | iwslt2017 | | |
| | ja | japanese | 10000 | iwslt2017 | | |
| | ko | korean | 10000 | iwslt2017 | | |
| | lt | lithuanian | 10000 | emea | | |
| | lv | latvian | 4595 | europa_ecdc_tm; europa_eac_tm | | |
| | mr | marathi | 10000 | tatoeba | | |
| | mt | maltese | 10000 | multi_para_crawl | | |
| | nl | dutch | 10000 | kde4 | | |
| | no | norwegian | 10000 | multi_para_crawl | | |
| | pl | polish | 10000 | ecb | | |
| | pt | portuguese | 10000 | tatoeba | | |
| | ro | romanian | 10000 | kde4 | | |
| | ru | russian | 10000 | xnli | | |
| | sk | slovak | 10000 | multi_para_crawl | | |
| | sl | slovenian | 4589 | europa_ecdc_tm; europa_eac_tm | | |
| | sw | swahili | 10000 | xnli | | |
| | sv | swedish | 10000 | kde4 | | |
| | th | thai | 10000 | xnli | | |
| | tl | tagalog | 10000 | multi_para_crawl | | |
| | tn | serpeti | 10000 | autshumato | | |
| | tr | turkish | 10000 | xnli | | |
| | ts | dzonga | 10000 | autshumato | | |
| | ur | urdu | 10000 | xnli | | |
| | vi | vietnamese | 10000 | xnli | | |
| | yo | yoruba | 9970 | menyo20k_mt | | |
| | zh | chinese | 10000 | xnli | | |
| | zu | zulu, south africa | 10000 | autshumato | | |
| """ | |
| def main(): | |
| args = get_args() | |
| subset_dataset_dict = dict() | |
| lines = s.strip().split("\n") | |
| with open(args.train_subset, "w", encoding="utf-8") as ftrain, open(args.valid_subset, "w", encoding="utf-8") as fvalid: | |
| for line in lines: | |
| row = str(line).split("|") | |
| row = [col.strip() for col in row if len(col) != 0] | |
| if len(row) != 4: | |
| raise AssertionError("not 4 item, line: {}".format(line)) | |
| abbr = row[0] | |
| full = row[1] | |
| total = int(row[2]) | |
| subsets = [e.strip() for e in row[3].split(";")] | |
| count = 0 | |
| for subset in subsets: | |
| if subset in subset_dataset_dict.keys(): | |
| dataset_dict = subset_dataset_dict[subset] | |
| else: | |
| dataset_dict = load_dataset( | |
| "qgyd2021/language_identification", | |
| name=subset, | |
| cache_dir=args.dataset_cache_dir, | |
| # download_mode=DownloadMode.FORCE_REDOWNLOAD | |
| ) | |
| subset_dataset_dict[subset] = dataset_dict | |
| train_dataset = dataset_dict["train"] | |
| for sample in train_dataset: | |
| text = sample["text"] | |
| language = sample["language"] | |
| data_source = sample["data_source"] | |
| if count > total: | |
| break | |
| if language != abbr: | |
| continue | |
| split = "train" if random.random() < 0.8 else "valid" | |
| row_ = { | |
| "text": text, | |
| "label": language, | |
| "language": full, | |
| "data_source": data_source, | |
| "split": split, | |
| } | |
| row_ = json.dumps(row_, ensure_ascii=False) | |
| if split == "train": | |
| ftrain.write("{}\n".format(row_)) | |
| elif split == "valid": | |
| fvalid.write("{}\n".format(row_)) | |
| else: | |
| raise AssertionError | |
| count += 1 | |
| return | |
| if __name__ == "__main__": | |
| main() | |