| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| import os |
|
|
| import datasets |
|
|
|
|
| _DESCRIPTION = """\ |
| Parallel corpora from Web Crawls collected in the ParaCrawl project. |
| |
| 42 languages, 43 bitexts |
| total number of files: 59,996 |
| total number of tokens: 56.11G |
| total number of sentence fragments: 3.13G |
| """ |
| _HOMEPAGE = "http://opus.nlpl.eu/ParaCrawl.php" |
| _CITATION = r"""\ |
| @inproceedings{banon-etal-2020-paracrawl, |
| title = "{P}ara{C}rawl: Web-Scale Acquisition of Parallel Corpora", |
| author = "Ba{\~n}{\'o}n, Marta and |
| Chen, Pinzhen and |
| Haddow, Barry and |
| Heafield, Kenneth and |
| Hoang, Hieu and |
| Espl{\`a}-Gomis, Miquel and |
| Forcada, Mikel L. and |
| Kamran, Amir and |
| Kirefu, Faheem and |
| Koehn, Philipp and |
| Ortiz Rojas, Sergio and |
| Pla Sempere, Leopoldo and |
| Ram{\'\i}rez-S{\'a}nchez, Gema and |
| Sarr{\'\i}as, Elsa and |
| Strelec, Marek and |
| Thompson, Brian and |
| Waites, William and |
| Wiggins, Dion and |
| Zaragoza, Jaume", |
| booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", |
| month = jul, |
| year = "2020", |
| address = "Online", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2020.acl-main.417", |
| doi = "10.18653/v1/2020.acl-main.417", |
| pages = "4555--4567", |
| } |
| @InProceedings{TIEDEMANN12.463, |
| author = {Jörg Tiedemann}, |
| title = {Parallel Data, Tools and Interfaces in OPUS}, |
| booktitle = {Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC'12)}, |
| year = {2012}, |
| month = {may}, |
| date = {23-25}, |
| address = {Istanbul, Turkey}, |
| editor = {Nicoletta Calzolari (Conference Chair) and Khalid Choukri and Thierry Declerck and Mehmet Uğur Doğan and Bente Maegaard and Joseph Mariani and Asuncion Moreno and Jan Odijk and Stelios Piperidis}, |
| publisher = {European Language Resources Association (ELRA)}, |
| isbn = {978-2-9517408-7-7}, |
| language = {english} |
| } |
| """ |
|
|
| _VERSION = "9.0.0" |
| _BASE_NAME = "ParaCrawl.{}.{}" |
| _BASE_URL = "https://object.pouta.csc.fi/OPUS-ParaCrawl/v9/moses/{}-{}.txt.zip" |
| |
| _LANGUAGE_PAIRS = [ |
| ("el", "en"), |
| ("en", "km"), |
| ("en", "so"), |
| ("de", "pl"), |
| ("fr", "nl"), |
| ("en", "sw"), |
| ("en", "tl"), |
| ("es", "gl"), |
| ] |
|
|
|
|
| class ParaCrawlConfig(datasets.BuilderConfig): |
| def __init__(self, *args, lang1=None, lang2=None, **kwargs): |
| super().__init__( |
| *args, |
| name=f"{lang1}-{lang2}", |
| **kwargs, |
| ) |
| assert lang1 != lang2, "'language 1' & 'language 2' should be different from each other" |
| self.lang1 = lang1 |
| self.lang2 = lang2 |
|
|
|
|
| class OpusParaCrawl(datasets.GeneratorBasedBuilder): |
| BUILDER_CONFIGS = [ |
| ParaCrawlConfig( |
| lang1=lang1, |
| lang2=lang2, |
| description=f"Translating {lang1} to {lang2} or vice versa", |
| version=datasets.Version(_VERSION), |
| ) |
| for lang1, lang2 in _LANGUAGE_PAIRS |
| ] |
| BUILDER_CONFIG_CLASS = ParaCrawlConfig |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "id": datasets.Value("string"), |
| "translation": datasets.Translation(languages=(self.config.lang1, self.config.lang2)), |
| }, |
| ), |
| supervised_keys=None, |
| homepage=_HOMEPAGE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| download_url = _BASE_URL.format(self.config.lang1, self.config.lang2) |
| path = dl_manager.download_and_extract(download_url) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={"datapath": path}, |
| ) |
| ] |
|
|
| def _generate_examples(self, datapath): |
| lang1, lang2 = self.config.lang1, self.config.lang2 |
| folder = lang1 + "-" + lang2 |
| lang1_filename = _BASE_NAME.format(folder, lang1) |
| lang2_filename = _BASE_NAME.format(folder, lang2) |
| lang1_path = os.path.join(datapath, lang1_filename) |
| lang2_path = os.path.join(datapath, lang2_filename) |
| with open(lang1_path, encoding="utf-8") as f1, open(lang2_path, encoding="utf-8") as f2: |
| for id_, (lang1_sentence, lang2_sentence) in enumerate(zip(f1, f2)): |
| lang1_sentence = lang1_sentence.strip() |
| lang2_sentence = lang2_sentence.strip() |
| yield id_, { |
| "id": str(id_), |
| "translation": {lang1: lang1_sentence, lang2: lang2_sentence}, |
| } |
|
|