Datasets:
License:
Commit
·
7d396d1
1
Parent(s):
15a95e7
Syncing the docs and the codes
Browse files- README.md +30 -6
- dedup_raw_wiki_data_indo.sh +2 -2
- extract_raw_wiki_data_indo.sh +1 -1
- indo_wiki.py +1 -1
README.md
CHANGED
|
@@ -33,9 +33,14 @@ task_ids:
|
|
| 33 |
pretty_name: Wikipedia Archive for Indonesian Languages & Local Languages
|
| 34 |
tags:
|
| 35 |
- Wikipedia
|
| 36 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
- Indonesian Language
|
| 38 |
-
-
|
| 39 |
- Indonesia-related Languages
|
| 40 |
- Indonesian Local Languages
|
| 41 |
dataset_info:
|
|
@@ -162,8 +167,27 @@ license: cc-by-sa-3.0
|
|
| 162 |
Welcome to Indonesian Wikipedia Data Repository. The datasets are extracted from [Wikipedia HF](https://huggingface.co/datasets/wikipedia) and processed using the scripts available in this repository for reproducibility purpose.
|
| 163 |
|
| 164 |
# **FAQS**
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 165 |
### How do I extract new Wikipedia Dataset of Indonesian languages?
|
| 166 |
-
You may check to the script [_```extract_raw_wiki_data.py```_](https://huggingface.co/datasets/sabilmakbar/
|
| 167 |
|
| 168 |
### How do I extract new Wikipedia Dataset of Indonesian languages?
|
| 169 |
You may visit this [Wikipedia Dump Index](https://dumps.wikimedia.org/backup-index.html) to check any latest available data and this link [Wikipedia Language Coverage](https://meta.wikimedia.org/wiki/List_of_Wikipedias#All_Wikipedias_ordered_by_number_of_articles) to map into any languages that you're wanting to extract.
|
|
@@ -171,7 +195,7 @@ You may visit this [Wikipedia Dump Index](https://dumps.wikimedia.org/backup-ind
|
|
| 171 |
### How does the data being preprocessed? What makes it different from loading it directly from Wikipedia HF?
|
| 172 |
The data available in here are processed with following flows:
|
| 173 |
1. Raw data is being deduplicated on ```title``` and ```text``` (text-content from a given article), to remove articles containing boilerplate text (template text that are used usually for no-available informations or asking for contributions of content in that article), which usually deemed noisy for NLP data.
|
| 174 |
-
2. Furthermore, the ```title``` and ```text``` data are being checked for string-matching duplication (duplication of text that are being pre-processed, i.e symbols removed, HTML tags striped, or ASCII-chars validated). You may check this [ ```cleanse_wiki_data.py```](https://huggingface.co/datasets/sabilmakbar/
|
| 175 |
|
| 176 |
# Getting Started #
|
| 177 |
### To read the datasets directly ###
|
|
@@ -179,14 +203,14 @@ Use one of the following code chunks to load it from HuggingFace Hub:
|
|
| 179 |
You can refer to the 2nd args of ```config name``` using the following script
|
| 180 |
```
|
| 181 |
dataset = load_dataset(
|
| 182 |
-
"sabilmakbar/
|
| 183 |
"indo_wiki_dedup_data" # a config name, can be "indo_wiki_raw_data" or "indowiki_dedup_id_only", defaults to "indo_wiki_dedup_data"
|
| 184 |
)
|
| 185 |
```
|
| 186 |
Or you can provide both ```lang``` and ```date_stamp``` (providing only one will thrown an error)
|
| 187 |
```
|
| 188 |
dataset = load_dataset(
|
| 189 |
-
"sabilmakbar/
|
| 190 |
lang = "id", # see the splits for complete lang choices
|
| 191 |
date_stamp="20230901"
|
| 192 |
)
|
|
|
|
| 33 |
pretty_name: Wikipedia Archive for Indonesian Languages & Local Languages
|
| 34 |
tags:
|
| 35 |
- Wikipedia
|
| 36 |
+
- Indonesian
|
| 37 |
+
- Sundanese
|
| 38 |
+
- Javanese
|
| 39 |
+
- Malay
|
| 40 |
+
- Dialect
|
| 41 |
+
- Javanese Dialect (Banyumase/Ngapak)
|
| 42 |
- Indonesian Language
|
| 43 |
+
- Malay Language
|
| 44 |
- Indonesia-related Languages
|
| 45 |
- Indonesian Local Languages
|
| 46 |
dataset_info:
|
|
|
|
| 167 |
Welcome to Indonesian Wikipedia Data Repository. The datasets are extracted from [Wikipedia HF](https://huggingface.co/datasets/wikipedia) and processed using the scripts available in this repository for reproducibility purpose.
|
| 168 |
|
| 169 |
# **FAQS**
|
| 170 |
+
### What are the available languages provided in dataset?
|
| 171 |
+
Please check the following table.
|
| 172 |
+
|Lang Code|Lang Desc|Wiki Info|Total Data|Total Size (bytes)|
|
| 173 |
+
|:---|:----:|:---:|
|
| 174 |
+
|ace|Acehnese|[Wiki Link](https://en.wikipedia.org/wiki/Acehnese_language)|12904|4867838|
|
| 175 |
+
|ban|Balinese|[Wiki Link](https://en.wikipedia.org/wiki/Balinese_language)|19837|17366080|
|
| 176 |
+
|bjn|Acehnese|[Wiki Link](https://en.wikipedia.org/wiki/Banjarese_language)|10437|6655378|
|
| 177 |
+
|bug|Buginese|[Wiki Link](https://en.wikipedia.org/wiki/Buginese_language)|9793|2072609|
|
| 178 |
+
|gor|Gorontalo|[Wiki Link](https://en.wikipedia.org/wiki/Gorontalo_language)|14514|5989252|
|
| 179 |
+
|id|Indonesian|[Wiki Link](https://en.wikipedia.org/wiki/Indonesian_language)|654287|1100932403|
|
| 180 |
+
|jv|Javanese|[Wiki Link](https://en.wikipedia.org/wiki/Javanese_language)|72667|69774853|
|
| 181 |
+
|map_bms|Banyumasan - Dialect of Javanese|[Wiki Link](https://en.wikipedia.org/wiki/Banyumasan_dialect)|11832|5060989|
|
| 182 |
+
|min|Minangkabau|[Wiki Link](https://en.wikipedia.org/wiki/Minangkabau_language)|225858|116376870|
|
| 183 |
+
|ms|Malay|[Wiki Link](https://en.wikipedia.org/wiki/Malay_language)|346186|410443550|
|
| 184 |
+
|nia|Nias|[Wiki Link](https://en.wikipedia.org/wiki/Nias_language)|1650|1938121|
|
| 185 |
+
|su|Sundanese|[Wiki Link](https://en.wikipedia.org/wiki/Sundanese_language)|61494|47410439|
|
| 186 |
+
|tet|Tetum|[Wiki Link](https://en.wikipedia.org/wiki/Tetum_language)|1465|1452716|
|
| 187 |
+
|
| 188 |
+
|
| 189 |
### How do I extract new Wikipedia Dataset of Indonesian languages?
|
| 190 |
+
You may check to the script [_```extract_raw_wiki_data.py```_](https://huggingface.co/datasets/sabilmakbar/indo_wiki/blob/main/extract_raw_wiki_data.py) to understand its implementations, or you can adjust the bash provided in [_```extract_raw_wiki_data_indo.sh```_](https://huggingface.co/datasets/sabilmakbar/indo_wiki/blob/main/extract_raw_wiki_data_indo.sh) to extract it on your own. Please note that this dataset is extensible to any languages of your choice.
|
| 191 |
|
| 192 |
### How do I extract new Wikipedia Dataset of Indonesian languages?
|
| 193 |
You may visit this [Wikipedia Dump Index](https://dumps.wikimedia.org/backup-index.html) to check any latest available data and this link [Wikipedia Language Coverage](https://meta.wikimedia.org/wiki/List_of_Wikipedias#All_Wikipedias_ordered_by_number_of_articles) to map into any languages that you're wanting to extract.
|
|
|
|
| 195 |
### How does the data being preprocessed? What makes it different from loading it directly from Wikipedia HF?
|
| 196 |
The data available in here are processed with following flows:
|
| 197 |
1. Raw data is being deduplicated on ```title``` and ```text``` (text-content from a given article), to remove articles containing boilerplate text (template text that are used usually for no-available informations or asking for contributions of content in that article), which usually deemed noisy for NLP data.
|
| 198 |
+
2. Furthermore, the ```title``` and ```text``` data are being checked for string-matching duplication (duplication of text that are being pre-processed, i.e symbols removed, HTML tags striped, or ASCII-chars validated). You may check this [ ```cleanse_wiki_data.py```](https://huggingface.co/datasets/sabilmakbar/indo_wiki/blob/main/cleanse_wiki_data.py) script to understand its implementation.
|
| 199 |
|
| 200 |
# Getting Started #
|
| 201 |
### To read the datasets directly ###
|
|
|
|
| 203 |
You can refer to the 2nd args of ```config name``` using the following script
|
| 204 |
```
|
| 205 |
dataset = load_dataset(
|
| 206 |
+
"sabilmakbar/indo_wiki",
|
| 207 |
"indo_wiki_dedup_data" # a config name, can be "indo_wiki_raw_data" or "indowiki_dedup_id_only", defaults to "indo_wiki_dedup_data"
|
| 208 |
)
|
| 209 |
```
|
| 210 |
Or you can provide both ```lang``` and ```date_stamp``` (providing only one will thrown an error)
|
| 211 |
```
|
| 212 |
dataset = load_dataset(
|
| 213 |
+
"sabilmakbar/indo_wiki",
|
| 214 |
lang = "id", # see the splits for complete lang choices
|
| 215 |
date_stamp="20230901"
|
| 216 |
)
|
dedup_raw_wiki_data_indo.sh
CHANGED
|
@@ -4,8 +4,8 @@
|
|
| 4 |
# "ace", "ban", "bjn", "bug", "gor", "id", "jv", "map-bms", "min", "ms", "nia", "su", "tet"
|
| 5 |
|
| 6 |
#params of executions
|
| 7 |
-
folder_dir_to_save=./
|
| 8 |
-
input_folder_to_be_dedup=./
|
| 9 |
|
| 10 |
drop_hard_dupl=True
|
| 11 |
drop_soft_dupl=True
|
|
|
|
| 4 |
# "ace", "ban", "bjn", "bug", "gor", "id", "jv", "map-bms", "min", "ms", "nia", "su", "tet"
|
| 5 |
|
| 6 |
#params of executions
|
| 7 |
+
folder_dir_to_save=./indo_wiki_dedup
|
| 8 |
+
input_folder_to_be_dedup=./indo_wiki_raw
|
| 9 |
|
| 10 |
drop_hard_dupl=True
|
| 11 |
drop_soft_dupl=True
|
extract_raw_wiki_data_indo.sh
CHANGED
|
@@ -5,7 +5,7 @@
|
|
| 5 |
|
| 6 |
#params of executions
|
| 7 |
date_ver=20230901
|
| 8 |
-
folder_dir_to_save=./
|
| 9 |
lang_list=(ace ban bjn)
|
| 10 |
# bug gor id jv map-bms min ms nia su tet)
|
| 11 |
|
|
|
|
| 5 |
|
| 6 |
#params of executions
|
| 7 |
date_ver=20230901
|
| 8 |
+
folder_dir_to_save=./indo_wiki_raw
|
| 9 |
lang_list=(ace ban bjn)
|
| 10 |
# bug gor id jv map-bms min ms nia su tet)
|
| 11 |
|
indo_wiki.py
CHANGED
|
@@ -18,7 +18,7 @@ _CITATIONS = """\
|
|
| 18 |
title = "Huggingface Wikipedia Dataset",
|
| 19 |
url = "https://huggingface.co/datasets/wikipedia"}"""
|
| 20 |
|
| 21 |
-
_REPO_URL = "https://huggingface.co/datasets/sabilmakbar/
|
| 22 |
|
| 23 |
_LICENSE = (
|
| 24 |
"This work is licensed under the Creative Commons Attribution-ShareAlike "
|
|
|
|
| 18 |
title = "Huggingface Wikipedia Dataset",
|
| 19 |
url = "https://huggingface.co/datasets/wikipedia"}"""
|
| 20 |
|
| 21 |
+
_REPO_URL = "https://huggingface.co/datasets/sabilmakbar/indo_wiki"
|
| 22 |
|
| 23 |
_LICENSE = (
|
| 24 |
"This work is licensed under the Creative Commons Attribution-ShareAlike "
|