File size: 863 Bytes
afae95f
7da30bf
afae95f
7da30bf
afae95f
 
7da30bf
 
 
 
afae95f
7da30bf
afae95f
 
7da30bf
afae95f
7da30bf
 
afae95f
7da30bf
 
 
 
 
 
afae95f
 
7da30bf
 
afae95f
7da30bf
afae95f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33

# English-Yoruba STEM Translation Model

This model is trained to translate English STEM content to Yoruba.

## Model Details
- **Architecture:** Transformer-based sequence-to-sequence model
- **Base Model:** Davlan/mt5-base-en-yor-mt
- **Training Data:** YorubaSTEM1.0
- **Performance:** BLEU: 36.08

## Usage


from transformers import AutoTokenizer, AutoModelForSeq2SeqLM

tokenizer = AutoTokenizer.from_pretrained("gbelewade/YorubaSTEMt5")
model = AutoModelForSeq2SeqLM.from_pretrained("gbelewade/YorubaSTEMt5")

# Translate English text to Yoruba
english_text = "The chemical formula for water is H2O."
inputs = tokenizer(english_text, return_tensors="pt")
outputs = model.generate(**inputs)
yoruba_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(yoruba_text)


## Limitations
[Describe any known limitations of the model]

## Citation