| from collections import defaultdict |
| from time import sleep |
|
|
| from datasets import DatasetDict, load_dataset |
| from huggingface_hub import HfApi, hf_hub_download |
|
|
| languages = [ |
| "de", |
| "bn", |
| "it", |
| "pt", |
| "nl", |
| "cs", |
| "ro", |
| "bg", |
| "sr", |
| "fi", |
| "fa", |
| "hi", |
| "da", |
| "en", |
| "no", |
| "sv", |
| ] |
|
|
|
|
| def map_corpus_to_query(example, negatives_dict): |
| query = example["query"] |
| title = example["title"] |
| positive = [example["text"]] |
| negatives = negatives_dict[title] |
| return {"query": query, "positive": positive, "negative": negatives} |
|
|
|
|
| ds_dict = DatasetDict() |
| for lang in languages: |
| sleep(5) |
| ds_queries = load_dataset( |
| f"rasdani/cohere-wikipedia-2023-11-{lang}-queries", split="train" |
| ) |
| ds_corpus = load_dataset( |
| f"rasdani/cohere-wikipedia-2023-11-{lang}-1.5k-articles", split="train" |
| ) |
| ds_corpus = ds_corpus.filter(lambda x: x["score"] != 1) |
| sleep(5) |
|
|
| negatives_dict = defaultdict(list) |
| for row in ds_corpus: |
| negatives_dict[row["title"]].append(row["text"]) |
|
|
| ds = ds_queries.map( |
| lambda x: map_corpus_to_query(x, negatives_dict), |
| remove_columns=ds_queries.column_names, |
| ) |
|
|
| repo_id = "ellamind/wikipedia-2023-11-reranking-multilingual" |
| ds.push_to_hub(repo_id, config_name=lang, split="test") |
|
|
| |
| sleep(5) |
| readme_path = hf_hub_download( |
| repo_id=repo_id, filename="README.md", repo_type="dataset" |
| ) |
|
|
| with open(readme_path, "r") as f: |
| readme_content = f.read() |
|
|
| readme = """ |
| This dataset is derived from Cohere's wikipedia-2023-11 dataset, which is in turn derived from `wikimedia/wikipedia`. |
| The dataset is licensed under the Creative Commons CC BY-SA 3.0 license. |
| """ |
| |
| if "- license: " not in readme_content and readme not in readme_content: |
| license = "cc-by-sa-3.0" |
|
|
| updated_readme = readme_content.replace( |
| "---\ndataset_info:", "---\nlicense: {license}\ndataset_info:" |
| ).format(license=license) |
| updated_readme += readme |
|
|
| api = HfApi() |
| readme_bytes = updated_readme.encode("utf-8") |
| api.upload_file( |
| path_or_fileobj=readme_bytes, |
| path_in_repo="README.md", |
| repo_id=repo_id, |
| repo_type="dataset", |
| ) |
|
|