File size: 1,287 Bytes
3befaa4
b023e22
 
 
 
 
 
 
 
3befaa4
 
b023e22
3befaa4
b023e22
3befaa4
b023e22
 
 
 
 
 
3befaa4
b023e22
 
 
 
 
 
 
3befaa4
b023e22
 
 
3befaa4
b023e22
 
3befaa4
b023e22
 
d2107c8
 
 
83f4eaa
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
tags:
- luganda
- iteso
- runyankore
- acholi
- ateso
- mistral-7b
- fine-tuned
---

# Alkebulan AI: Fine-Tuned Mistral-7B for Ugandan Languages

This is a fine-tuned version of the Mistral-7B model for **Luganda**, **Iteso**, **Runyankore**, **Acholi**, and **Ateso** languages. The model is trained on **parallel datasets** to enable translation and basic interaction in these languages.

## Languages Supported
- Luganda
- Iteso
- Runyankore
- Acholi
- Ateso

## Training Data
The model was fine-tuned on parallel datasets containing:
- English to Luganda
- English to Iteso
- English to Runyankore
- English to Acholi
- English to Ateso

## Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("Psalms23Wave/Alkebulan-AI")
model = AutoModelForCausalLM.from_pretrained("Psalms23Wave/Alkebulan-AI")

inputs = tokenizer("Translate to Luganda: Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

## Disclaimer
This model is licensed under the CC BY-NC-ND 4.0 license. 
It is intended for **non-commercial use only. 
You may not use this model for commercial purposes, 
and you may not distribute modified versions of the model.