Datasets:
Update README.md
Browse files
README.md
CHANGED
|
@@ -7868,6 +7868,20 @@ configs:
|
|
| 7868 |
data_files:
|
| 7869 |
- split: label
|
| 7870 |
path: zu/label-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7871 |
---
|
| 7872 |
|
| 7873 |
Entity label data extracted from Wikidata (wikidata-20220103-all.json.gz), filtered for item entities only
|
|
@@ -7879,7 +7893,7 @@ You can generate parallel text examples from this dataset like below:
|
|
| 7879 |
from datasets import load_dataset
|
| 7880 |
import pandas as pd
|
| 7881 |
|
| 7882 |
-
def
|
| 7883 |
out_df = None
|
| 7884 |
for lc in lang_codes:
|
| 7885 |
dataset = load_dataset(repo_id, lc, **datasets_config)
|
|
@@ -7895,5 +7909,6 @@ def get_lang_pair(lang_codes: list, how="inner", repo_id="rayliuca/wikidata_enti
|
|
| 7895 |
)
|
| 7896 |
return out_df
|
| 7897 |
|
| 7898 |
-
|
| 7899 |
-
|
|
|
|
|
|
| 7868 |
data_files:
|
| 7869 |
- split: label
|
| 7870 |
path: zu/label-*
|
| 7871 |
+
task_categories:
|
| 7872 |
+
- translation
|
| 7873 |
+
- text2text-generation
|
| 7874 |
+
language:
|
| 7875 |
+
- en
|
| 7876 |
+
- fr
|
| 7877 |
+
- de
|
| 7878 |
+
- ja
|
| 7879 |
+
- zh
|
| 7880 |
+
- hi
|
| 7881 |
+
- ar
|
| 7882 |
+
- bn
|
| 7883 |
+
- ru
|
| 7884 |
+
- es
|
| 7885 |
---
|
| 7886 |
|
| 7887 |
Entity label data extracted from Wikidata (wikidata-20220103-all.json.gz), filtered for item entities only
|
|
|
|
| 7893 |
from datasets import load_dataset
|
| 7894 |
import pandas as pd
|
| 7895 |
|
| 7896 |
+
def parallel_labels(lang_codes: list, how="inner", repo_id="rayliuca/wikidata_entity_label", merge_config={}, datasets_config={}) -> pd.DataFrame:
|
| 7897 |
out_df = None
|
| 7898 |
for lc in lang_codes:
|
| 7899 |
dataset = load_dataset(repo_id, lc, **datasets_config)
|
|
|
|
| 7909 |
)
|
| 7910 |
return out_df
|
| 7911 |
|
| 7912 |
+
# Note: the "en" subset is >4GB
|
| 7913 |
+
parallel_labels(['en', 'fr', 'ja', 'zh'])
|
| 7914 |
+
```
|