M2M-100 Doc-clean โ Hieroglyphic-to-German Translation
This model is part of the paper "Data Contamination in Neural Machine Translation of Ancient Egyptian Hieroglyphics" (NLP4DH 2026).
Model Description
M2M-100 continued fine-tuning from Conservative checkpoint on document-level filtered training data (13,365 samples, removing all training sentences from test-source documents). Best checkpoint at step 300.
Task: Ancient Egyptian hieroglyphics (Gardiner notation) โ German translation
Performance
| Subset | BLEU |
|---|---|
| All (n=50) | 45.2 |
| Contaminated (n=16) | 73.3 |
| Clean (n=34) | 40.3 |
Important: The "All" and "Contaminated" BLEU scores are inflated due to target-side data contamination (32% of test targets appear in training). The Clean score represents genuine translation quality on decontaminated samples.
Usage
from transformers import M2MForConditionalGeneration, M2MTokenizer
model = M2MForConditionalGeneration.from_pretrained("bumblelbee/hiero-m2m100-doc-clean")
tokenizer = M2MTokenizer.from_pretrained("bumblelbee/hiero-m2m100-doc-clean")
# Gardiner notation input (hieroglyphic transliteration)
source = "D36 N35 G17 D21 X1 O34"
tokenizer.src_lang = "ea"
inputs = tokenizer(source, return_tensors="pt")
generated = model.generate(**inputs, forced_bos_token_id=tokenizer.get_lang_id("de"))
output = tokenizer.decode(generated[0], skip_special_tokens=True)
print(output)
Training Data
Fine-tuned on 18,669 eaโde pairs from the Thesaurus Linguae Aegyptiae (TLA), maintained by the Berlin-Brandenburg Academy of Sciences and Humanities.
Citation
@inproceedings{contamination2026nlp4dh,
title={Data Contamination in Neural Machine Translation of Ancient Egyptian Hieroglyphics},
booktitle={Proceedings of the Workshop on Natural Language Processing for Digital Humanities (NLP4DH 2026)},
year={2026}
}
Paper Repository
See the full paper, scripts, and results: GitHub repository
- Downloads last month
- 13
Model tree for bumblelbee/hiero-m2m100-doc-clean
Base model
facebook/m2m100_418M