File size: 2,259 Bytes
4fedda1
 
 
 
 
 
 
 
 
 
 
bba63da
4fedda1
bba63da
 
4fedda1
 
 
 
bba63da
4fedda1
bba63da
4fedda1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bba63da
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4fedda1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
---
language:
- ar
- en
tags:
- chat
- translation
license: apache-2.0
base_model: OMDA-Decoder
pipeline_tag: text-generation
---

# OMDA: Bilingual Arabic-English Chat LLM

**Model Name:** OMDA  
**Architecture:** OMDA-Decoder (Custom Architecture)  
**Tokenizer:** OMDATokenizer (Custom Tokenizer)  
**Languages:** Arabic (Primary), English  
**Model Type:** Chat/Instruction-following  
**Author:** Binomda  
**Release Date:** 2025-06-28  

## Model Overview
OMDA is a compact bilingual language model specifically designed for Arabic-English conversational AI applications. Built with a custom decoder architecture, it excels at understanding and generating natural responses in both languages.

## Model Specifications
| Parameter          | Value     |
|--------------------|-----------|
| Layers             | 6         |
| Hidden Size        | 512       |
| Attention Heads    | 8         |
| FFN Dimension      | 2048      |
| Max Sequence Length| 512       |
| Vocabulary Size    | 128,004   |
| Training Data      | 1,000 curated Arabic-English conversation pairs |

## Intended Uses
โœ… Chatbot development  
โœ… Bilingual assistant applications  
โœ… Educational tools  
โœ… Basic translation tasks  

## Quick Start
```python
from transformers import pipeline

# Initialize chat pipeline
chatbot = pipeline("text-generation", model="BINOMDA/OMDA")

# Arabic input example
ar_responser = chatbot("ู…ุง ู‡ูˆ ุฑุฃูŠูƒ ููŠ ุงู„ุชูƒู†ูˆู„ูˆุฌูŠุง ุงู„ุญุฏูŠุซุฉุŸ")

# English input example
en_response = chatbot("Explain artificial intelligence simply")

## Intended Use
- Chatbots, assistants, translation, and educational tools for Arabic/English.

## Training
- Trained for 5 epochs on 1000 samples.
- Loss curve and checkpoints included.

## Limitations
- This is a small-scale demonstration model and may not generalize well to all real-world chat scenarios.
- Not suitable for production use without further scaling, extensive evaluation, and safety checks.
- Limited training data and model size may result in hallucinations or inaccurate translations.
- No advanced filtering for inappropriate or biased outputs.
- For research and educational purposes only.

## Export & Deployment
- See below for HuggingFace, llama.cpp, and ollama export