|
|
--- |
|
|
pretty_name: FormosanBank Machine Translation |
|
|
license: cc-by-4.0 |
|
|
task_categories: |
|
|
- translation |
|
|
language: |
|
|
|
|
|
- ami |
|
|
- bnn |
|
|
- ckv |
|
|
- dru |
|
|
- pwn |
|
|
- pyu |
|
|
- ssf |
|
|
- sxr |
|
|
- szy |
|
|
- tao |
|
|
- tay |
|
|
- trv |
|
|
- tsu |
|
|
- xnb |
|
|
- xsy |
|
|
|
|
|
- en |
|
|
- zh |
|
|
size_categories: |
|
|
- 100K<n<1M |
|
|
tags: |
|
|
- translation |
|
|
- machine-translation |
|
|
- low-resource |
|
|
- endangered-languages |
|
|
- formosan-languages |
|
|
- text |
|
|
library_name: datasets |
|
|
configs: |
|
|
- config_name: formosan-en |
|
|
data_files: "formosan_en_hf.csv" |
|
|
- config_name: formosan-zh |
|
|
data_files: "formosan_zh_hf.csv" |
|
|
|
|
|
--- |
|
|
# FormosanBank Machine Translation |
|
|
|
|
|
Parallel corpora for 15 Indigenous Formosan languages aligned to English and Mandarin Chinese, prepared for use with the Hugging Face `datasets` library. |
|
|
|
|
|
The dataset aggregates processed sentence- and phrase-level corpora into two CSV files: |
|
|
|
|
|
- **Formosan → English** (`formosan_en_hf.csv`) |
|
|
- **Formosan → Chinese** (`formosan_zh_hf.csv`) |
|
|
|
|
|
Each row is a single bilingual sentence pair with language, dialect, split, and provenance metadata. The dataset is designed for training and evaluating neural machine translation (NMT) and related models for low-resource Formosan languages. |
|
|
|
|
|
> **IMPORTANT DISCLAIMER:** |
|
|
> Our Machine Translation models published on HuggingFace and in our papers were trained on this data in addition to private data not available to the public due to content restrictions. |
|
|
> |
|
|
--- |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
- **Total sentence pairs:** 393,634 |
|
|
- **Formosan → English:** 85,144 |
|
|
- **Formosan → Chinese:** 308,490 |
|
|
- **Languages (15):** Amis, Bunun, Kavalan, Rukai, Paiwan, Puyuma, Thao, Saaroa, Sakizaya, Yami/Tao, Atayal, Seediq/Truku, Tsou, Kanakanavu, Saisiyat |
|
|
- **Targets:** English (`en`), Mandarin Chinese (`zh`) |
|
|
- **Splits (all languages, both targets combined):** |
|
|
- Train: 334,772 |
|
|
- Validate: 29,412 |
|
|
- Test: 29,450 |
|
|
- **License:** CC BY 4.0 |
|
|
- **Format:** UTF-8 CSV, one sentence pair per row |
|
|
|
|
|
The dataset is intended to support research on low-resource MT, cross-lingual transfer, and documentation of endangered Formosan languages. |
|
|
|
|
|
--- |
|
|
|
|
|
## Supported Tasks and Use Cases |
|
|
|
|
|
**Primary task** |
|
|
|
|
|
- `translation` |
|
|
- Formosan language → English |
|
|
- Formosan language → Chinese |
|
|
|
|
|
**Example use cases** |
|
|
|
|
|
- Training NMT systems (e.g. NLLB / encoder–decoder models) for individual Formosan languages. |
|
|
- Cross-lingual pretraining and evaluation for multilingual models. |
|
|
- Dialect-aware MT experiments using the `dialect` field. |
|
|
- Lexicon / dictionary-style MT from short phrases and headwords. |
|
|
|
|
|
--- |
|
|
|
|
|
## Languages and Coverage |
|
|
|
|
|
High-level sentence counts per language (summing both directions: Formosan→English and Formosan→Chinese): |
|
|
|
|
|
| Language | Formosan→English | Formosan→Chinese | Total | |
|
|
|---------------|------------------|------------------|--------| |
|
|
| Amis | 10,523 | 30,646 | 41,169 | |
|
|
| Bunun | 9,006 | 30,878 | 39,884 | |
|
|
| Kavalan | 2,098 | 14,682 | 16,780 | |
|
|
| Rukai | 11,850 | 39,360 | 51,210 | |
|
|
| Paiwan | 9,806 | 24,015 | 33,821 | |
|
|
| Puyuma | 7,199 | 26,154 | 33,353 | |
|
|
| Thao | 2,086 | 11,633 | 13,719 | |
|
|
| Saaroa | 2,130 | 9,819 | 11,949 | |
|
|
| Sakizaya | 2,132 | 11,318 | 13,450 | |
|
|
| Yami/Tao | 3,009 | 12,792 | 15,801 | |
|
|
| Atayal | 11,724 | 35,471 | 47,195 | |
|
|
| Seediq/Truku | 7,244 | 29,840 | 37,084 | |
|
|
| Tsou | 2,117 | 8,861 | 10,978 | |
|
|
| Kanakanavu | 2,105 | 11,904 | 14,009 | |
|
|
| Saisiyat | 2,115 | 11,117 | 13,232 | |
|
|
| **TOTAL** | **85,144** | **308,490** | **393,634** | |
|
|
|
|
|
Many languages also include **dialect labels**, for example: |
|
|
|
|
|
- Amis: UNKNOWN, Southern, Malan, Coastal, Xiuguluan, Hengchun |
|
|
- Bunun: UNKNOWN, Junqun, Luanqun, Kaqun, Tanqun, Zhuoqun |
|
|
- Paiwan, Puyuma, Rukai, Atayal, Seediq/Truku: multiple dialects |
|
|
- Others (e.g. Kavalan, Thao, Saaroa, Tsou, Kanakanavu, Saisiyat, Sakizaya, Yami/Tao) currently use `UNKNOWN` dialect |
|
|
|
|
|
Dialect coverage makes it possible to do dialect-specific MT or robustness studies. |
|
|
|
|
|
--- |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
### Data Files |
|
|
|
|
|
- `formosan_en_hf.csv` – all Formosan→English pairs |
|
|
- `formosan_zh_hf.csv` – all Formosan→Chinese pairs |
|
|
|
|
|
Each file contains all languages and splits. The **language direction** and **split** are specified per row. |
|
|
|
|
|
### Data Fields |
|
|
|
|
|
All CSVs share the same schema: |
|
|
|
|
|
```text |
|
|
id,source_lang,target_lang,source_sentence,target_sentence,lang_code,dialect,source,split |
|
|
```` |
|
|
|
|
|
* `id` *(int)* – unique row identifier within each file. |
|
|
* `source_lang` *(str)* – language code of the Formosan language (e.g. `"ami"`, `"bnn"`). |
|
|
* `target_lang` *(str)* – target language code (`"en"` or `"zh"`). |
|
|
* `source_sentence` *(str)* – sentence or phrase in the Formosan language. |
|
|
* `target_sentence` *(str)* – translation into the target language. |
|
|
* `lang_code` *(str)* – canonical code for the Formosan language (usually same as `source_lang`). |
|
|
* `dialect` *(str)* – dialect label (e.g. `"Southern"`, `"Malan"`, `"UNKNOWN"`). |
|
|
* `source` *(str)* – provenance string or original file path in the upstream corpora. |
|
|
* `split` *(str)* – one of `"train"`, `"validate"`, `"test"`. |
|
|
|
|
|
### Splits |
|
|
|
|
|
Splits are defined **per row** via the `split` column: |
|
|
|
|
|
* `train` – training data |
|
|
* `validate` – development / validate data |
|
|
* `test` – held-out test data |
|
|
|
|
|
Global totals across all languages and directions: |
|
|
|
|
|
* Train: 334,772 |
|
|
* Validate: 29,412 |
|
|
* Test: 29,450 |
|
|
|
|
|
Users can filter to any language pair and then re-group into a `DatasetDict` by `split`. |
|
|
|
|
|
--- |
|
|
|
|
|
## How to Load the Dataset |
|
|
|
|
|
### 1. Install dependencies |
|
|
|
|
|
```bash |
|
|
pip install datasets |
|
|
# optional, if you plan to fine-tune models: |
|
|
pip install transformers |
|
|
``` |
|
|
|
|
|
### 2. Load the EN and ZH files from the Hub |
|
|
|
|
|
Assume the dataset identifier is: |
|
|
|
|
|
```text |
|
|
FormosanBankDemos/formosan-mt |
|
|
``` |
|
|
|
|
|
Load both CSVs: |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
HF_ID = "FormosanBankDemos/formosan-mt" |
|
|
|
|
|
# Formosan → English |
|
|
ds_en_all = load_dataset( |
|
|
HF_ID, |
|
|
data_files="formosan_en_hf.csv", |
|
|
)["train"] # entire CSV exposed as a 'train' split by default |
|
|
|
|
|
# Formosan → Chinese |
|
|
ds_zh_all = load_dataset( |
|
|
HF_ID, |
|
|
data_files="formosan_zh_hf.csv", |
|
|
)["train"] |
|
|
``` |
|
|
|
|
|
Alternatively, if you rely on the YAML `configs` defined above: |
|
|
|
|
|
```python |
|
|
# Uses config_name: "formosan-en" from the README metadata |
|
|
ds_en_all = load_dataset( |
|
|
HF_ID, |
|
|
name="formosan-en", |
|
|
split="train", |
|
|
) |
|
|
``` |
|
|
|
|
|
### 3. Filter to a specific language pair (example: Amis → English, `ami → en`) |
|
|
|
|
|
```python |
|
|
ami_en = ds_en_all.filter( |
|
|
lambda ex: ex["source_lang"] == "ami" and ex["target_lang"] == "en" |
|
|
) |
|
|
|
|
|
print(ami_en) |
|
|
# Dataset({ |
|
|
# features: ['id', 'source_lang', 'target_lang', 'source_sentence', ...], |
|
|
# num_rows: ... |
|
|
# }) |
|
|
``` |
|
|
|
|
|
### 4. Get train / validation / test splits |
|
|
|
|
|
```python |
|
|
from datasets import DatasetDict |
|
|
|
|
|
def split_by_column(ds): |
|
|
return DatasetDict({ |
|
|
"train": ds.filter(lambda ex: ex["split"] == "train"), |
|
|
"validate": ds.filter(lambda ex: ex["split"] == "validate"), |
|
|
"test": ds.filter(lambda ex: ex["split"] == "test"), |
|
|
}) |
|
|
|
|
|
ami_en_splits = split_by_column(ami_en) |
|
|
|
|
|
print(ami_en_splits) |
|
|
# DatasetDict({ |
|
|
# train: Dataset({ ... }) |
|
|
# validate: Dataset({ ... }) |
|
|
# test: Dataset({ ... }) |
|
|
# }) |
|
|
``` |
|
|
|
|
|
### 5. (Optional) Add a `translation` column |
|
|
|
|
|
Many translation training scripts expect a `translation` field like `{"ami": "...", "en": "..."}`. You can construct it from existing columns: |
|
|
|
|
|
```python |
|
|
def add_translation(batch): |
|
|
translations = [] |
|
|
for src, tgt, sl, tl in zip( |
|
|
batch["source_sentence"], |
|
|
batch["target_sentence"], |
|
|
batch["source_lang"], |
|
|
batch["target_lang"], |
|
|
): |
|
|
translations.append({sl: src, tl: tgt}) |
|
|
return {"translation": translations} |
|
|
|
|
|
ami_en_splits = ami_en_splits.map(add_translation, batched=True) |
|
|
|
|
|
print(ami_en_splits["train"][0]["translation"]) |
|
|
# {'ami': "sa'osi", 'en': 'true'} |
|
|
``` |
|
|
|
|
|
You can reuse the same pattern for any other language pair: |
|
|
|
|
|
```python |
|
|
# Example: Paiwan → English |
|
|
pwn_en = ds_en_all.filter( |
|
|
lambda ex: ex["source_lang"] == "pwn" and ex["target_lang"] == "en" |
|
|
) |
|
|
pwn_en_splits = split_by_column(pwn_en) |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## Intended Uses, Limitations, and Risks |
|
|
|
|
|
### Intended Uses |
|
|
|
|
|
* Research on **low-resource machine translation** for Formosan languages. |
|
|
* Studies of **dialect variation** in MT via the `dialect` field. |
|
|
* Baseline and benchmark datasets for multilingual models focusing on Austronesian languages. |
|
|
|
|
|
### Limitations |
|
|
|
|
|
* Domain coverage is heterogeneous (dictionary-style entries, short phrases, and some longer sentences); performance may not generalize to all real-world text genres. |
|
|
* Dialect labels are not always available; some corpora use `UNKNOWN` for dialect. |
|
|
* The dataset currently encodes translations only **into** English and Chinese, not between Formosan languages. |
|
|
|
|
|
### Risks and Biases |
|
|
|
|
|
* Source corpora may contain historical, religious, or culturally specific content that is not representative of contemporary language use. |
|
|
* Translations may include inconsistencies or legacy orthography; users should verify quality before high-stakes use. |
|
|
* As with any MT dataset for endangered languages, there is a risk of misinterpretation or over-reliance on automatically produced translations in sensitive cultural contexts. |
|
|
|
|
|
Users should avoid deploying models trained on this dataset in critical or high-stakes settings without human expert review. |
|
|
|
|
|
--- |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this dataset in academic work, please cite the FormosanBank project and this dataset page. A generic citation format is: |
|
|
|
|
|
FormosanBank annotations and metadata are CC-BY-4.0. This means you must cite the source in any redistributed or derived products. |
|
|
For code packages, you may refer to the GitHub repository. For academic publications, you should cite Mohamed, W., Le Ferrand, É., Sung, L.-M., Prud'hommeaux, E., & Hartshorne, J. K. (2024). |
|
|
FormosanBank. Electronic Resource. |
|
|
|
|
|
> FormosanBankDemos. *FormosanBank Machine Translation Dataset*. Hugging Face Datasets. |
|
|
> Available at: [https://huggingface.co/datasets/FormosanBankDemos/formosan-mt](https://huggingface.co/datasets/FormosanBankDemos/formosan-mt) |
|
|
|
|
|
|
|
|
|