M2M-100 Doc-clean โ€” Hieroglyphic-to-German Translation

This model is part of the paper "Data Contamination in Neural Machine Translation of Ancient Egyptian Hieroglyphics" (NLP4DH 2026).

Model Description

M2M-100 continued fine-tuning from Conservative checkpoint on document-level filtered training data (13,365 samples, removing all training sentences from test-source documents). Best checkpoint at step 300.

Task: Ancient Egyptian hieroglyphics (Gardiner notation) โ†’ German translation

Performance

Subset BLEU
All (n=50) 45.2
Contaminated (n=16) 73.3
Clean (n=34) 40.3

Important: The "All" and "Contaminated" BLEU scores are inflated due to target-side data contamination (32% of test targets appear in training). The Clean score represents genuine translation quality on decontaminated samples.

Usage

from transformers import M2MForConditionalGeneration, M2MTokenizer

model = M2MForConditionalGeneration.from_pretrained("bumblelbee/hiero-m2m100-doc-clean")
tokenizer = M2MTokenizer.from_pretrained("bumblelbee/hiero-m2m100-doc-clean")

# Gardiner notation input (hieroglyphic transliteration)
source = "D36 N35 G17 D21 X1 O34"

tokenizer.src_lang = "ea"
inputs = tokenizer(source, return_tensors="pt")
generated = model.generate(**inputs, forced_bos_token_id=tokenizer.get_lang_id("de"))
output = tokenizer.decode(generated[0], skip_special_tokens=True)
print(output)

Training Data

Fine-tuned on 18,669 eaโ†’de pairs from the Thesaurus Linguae Aegyptiae (TLA), maintained by the Berlin-Brandenburg Academy of Sciences and Humanities.

Citation

@inproceedings{contamination2026nlp4dh,
  title={Data Contamination in Neural Machine Translation of Ancient Egyptian Hieroglyphics},
  booktitle={Proceedings of the Workshop on Natural Language Processing for Digital Humanities (NLP4DH 2026)},
  year={2026}
}

Paper Repository

See the full paper, scripts, and results: GitHub repository

Downloads last month
13
Safetensors
Model size
0.5B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for bumblelbee/hiero-m2m100-doc-clean

Finetuned
(156)
this model