nllb-lua-en-mt-v1

This model is a bidirectional English (eng) ↔ Tshiluba (lua) translation model. It is a fine-tuned version of SalomonMetre13/nllb-lua-en-mt-v1, specifically optimized for translation in the Tshiluba language context.

Model Description

  • Developed by: Salomon Metre
  • Model Type: NLLB (No Language Left Behind) Encoder-Decoder
  • Language(s): English (eng_Latn), Tshiluba (lua_Latn)
  • License: CC-BY-NC-4.0
  • Fine-tuned from: facebook/nllb-200-distilled-600M

Training and Evaluation Data

The model was fine-tuned on a parallel corpus of scraped Bible-based sentences. This dataset provides a critical foundation for Tshiluba, a low-resource language with limited digital parallel corpora.

Intended Uses & Limitations

Intended Use

  • Research on machine translation for Congolese/Bantu languages.
  • Practical drafting of translations between English and Tshiluba.

Limitations

  • Domain Specificity: Performance is strongest on formal or scriptural text and may decrease on colloquial or highly technical English/Tshiluba.
  • Morphological Complexity: As Tshiluba is a Bantu language with complex agglutinative morphology, the model may occasionally struggle with specific prefix/suffix agreements in out-of-distribution sentences.

Training Procedure

Training Hyperparameters

The following hyperparameters were used during training:

  • Learning Rate: 3e-05
  • Train Batch Size: 4
  • Eval Batch Size: 4
  • Optimizer: AdamW (Fused)
  • LR Scheduler: Linear with 200 warmup steps
  • Mixed Precision: Native AMP (FP16)

Evaluation Results (at step 4000)

  • Eval Loss: 0.1439
  • Epoch: 0.29 (partial epoch)

Framework Versions

  • Transformers: 4.51.3
  • Pytorch: 2.6.0+cu124
  • Datasets: 3.6.0
  • Tokenizers: 0.21.1
Downloads last month
10
Safetensors
Model size
0.6B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for SalomonMetre13/nllb-lua-en-mt-v1

Finetuned
(238)
this model