File size: 2,474 Bytes
666a104
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
# Machine Translation Model: English ↔ Tigrinya

This model is a fine-tuned machine translation model trained to translate between English and Tigrinya. It was trained on the parallel corpus of English and Tigrinya sentences.

## Model Overview

- **Model Type**: MarianMT (Multilingual Transformer Model)
- **Languages**: English ↔ Tigrinya
- **Model Architecture**: MarianMT, fine-tuned for English ↔ Tigrinya translation
- **Training Framework**: Hugging Face Transformers, PyTorch

## Training Details

- **Training Dataset**: NLLB Parallel Corpus (English ↔ Tigrinya)
- **Training Epochs**: 3
- **Batch Size**: 8
- **Max Length**: 128 tokens
- **Learning Rate**: Starts from `1.44e-07` and decays during training
- **Training Loss**: 
    - Final training loss: 0.4756
    - Per-epoch loss progress:
      - Epoch 1: 0.443
      - Epoch 2: 0.4077
      - Epoch 3: 0.4379

- **Gradient Norms**: 
    - Epoch 1: 1.14
    - Epoch 2: 1.11
    - Epoch 3: 1.06

- **Training Time**: 43376.7 seconds (~12 hours)
- **Training Speed**:
    - Training samples per second: 96.7
    - Training steps per second: 12.08

## Model Usage

This model can be used for translating English sentences to Tigrinya and vice versa.

### Example Usage (Python)

```python
from transformers import MarianMTModel, MarianTokenizer

# Load the model and tokenizer
model_name = "Hailay/MachineT_TigEng"  
model = MarianMTModel.from_pretrained(model_name)
tokenizer = MarianTokenizer.from_pretrained(model_name)

# Translate an English sentence to Tigrinya
english_text = "We must obey the Lord and leave them alone"
encoded_input = tokenizer(english_text, return_tensors="pt", padding=True, truncation=True)
translated = model.generate(**encoded_input)
translated_text = tokenizer.decode(translated[0], skip_special_tokens=True)

print(f"Translated text: {translated_text}")

##  Model Card

This model is trained to handle general English to Tigrinya translation tasks. It is suitable for a wide range of text, but might not perform well on domain-specific language or specialized terminology unless fine-tuned further.

##Model Architecture
The model is based on the MarianMT architecture, a transformer model designed for multilingual machine translation. It has been fine-tuned on English ↔ Tigrinya data.
##Acknowledgements
Corpus Name: NLLB

Package: NLLB.am-en in Moses format

Website: NLLB Corpus

If you use this model or the NLLB corpus in your work, please cite it as follows: